aboutsummaryrefslogtreecommitdiff
path: root/exec/java-exec/src/main/java/org/apache
diff options
context:
space:
mode:
authorTimothy Farkas <timothyfarkas@apache.org>2018-08-01 12:32:36 -0700
committerVitalii Diravka <vitalii.diravka@gmail.com>2018-08-10 21:01:15 +0300
commit3e3cd2bb76f6a98ca50798bdaa2d86777c7873ad (patch)
tree817f3e19414b4ec98c06ffea31d6a08227156fa8 /exec/java-exec/src/main/java/org/apache
parent9a90164ceedc65f33e3a01d3e55282d19de8009b (diff)
DRILL-6656: Disallow extra semicolons and multiple statements on the same line.
closes #1415
Diffstat (limited to 'exec/java-exec/src/main/java/org/apache')
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java3
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java6
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java6
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java4
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java6
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java18
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java3
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java59
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java4
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java3
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java8
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java4
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java4
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java3
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java3
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java4
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java6
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java35
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java4
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java4
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java4
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java3
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java100
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java5
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java2
-rw-r--r--exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java2
67 files changed, 233 insertions, 152 deletions
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java b/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
index f5556cf9b..9bec3933a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java
@@ -49,7 +49,7 @@ public final class ExecConstants {
public static final String ZK_REFRESH = "drill.exec.zk.refresh";
public static final String BIT_RETRY_TIMES = "drill.exec.rpc.bit.server.retry.count";
public static final String BIT_RETRY_DELAY = "drill.exec.rpc.bit.server.retry.delay";
- public static final String BIT_TIMEOUT = "drill.exec.bit.timeout" ;
+ public static final String BIT_TIMEOUT = "drill.exec.bit.timeout";
public static final String SERVICE_NAME = "drill.exec.cluster-id";
public static final String INITIAL_BIT_PORT = "drill.exec.rpc.bit.server.port";
public static final String INITIAL_DATA_PORT = "drill.exec.rpc.bit.server.dataport";
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java b/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java
index 84c37ba47..8fcb9e921 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java
@@ -29,8 +29,7 @@ public enum SerializationDefinition {
STORAGE_PLUGINS(3003, StoragePlugins.class),
FRAGMENT_STATUS(3004, FragmentStatus.class),
FRAGMENT_HANDLE(3005, FragmentHandle.class),
- PLAN_FRAGMENT(3006, PlanFragment.class)
- ;
+ PLAN_FRAGMENT(3006, PlanFragment.class);
public final int id;
public final Class<?> clazz;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java
index ccd7e3636..bca59dc2e 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java
@@ -824,7 +824,7 @@ public class DrillClient implements Closeable, ConnectionThrottle {
private class ListHoldingResultsListener implements UserResultsListener {
private final Vector<QueryDataBatch> results = new Vector<>();
private final SettableFuture<List<QueryDataBatch>> future = SettableFuture.create();
- private final UserProtos.RunQuery query ;
+ private final UserProtos.RunQuery query;
public ListHoldingResultsListener(UserProtos.RunQuery query) {
logger.debug( "Listener created for query \"\"\"{}\"\"\"", query );
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java
index ff0d0b247..7458c22dd 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java
@@ -180,7 +180,7 @@ public class DumpCat {
aggBatchMetaInfo.add(getBatchMetaInfo(vcSerializable));
if (vectorContainer.getRecordCount() == 0) {
- emptyBatchNum ++;
+ emptyBatchNum++;
}
if (prevSchema != null && !vectorContainer.getSchema().equals(prevSchema)) {
@@ -188,7 +188,7 @@ public class DumpCat {
}
prevSchema = vectorContainer.getSchema();
- batchNum ++;
+ batchNum++;
vectorContainer.zeroVectors();
}
@@ -219,7 +219,7 @@ public class DumpCat {
VectorAccessibleSerializable vcSerializable = null;
- while (input.available() > 0 && batchNum ++ < targetBatchNum) {
+ while (input.available() > 0 && batchNum++ < targetBatchNum) {
vcSerializable = new VectorAccessibleSerializable(DumpCat.allocator);
vcSerializable.readFromStream(input);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java b/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java
index ebb31f492..f15c01a02 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java
@@ -58,7 +58,7 @@ public class CodeGeneratorMethod implements Iterable<CodeGeneratorArgument> {
throw new RuntimeException(String.format("Unexpected number of parameter names %s. Expected %s on method %s.", Arrays.toString(parameterNames), Arrays.toString(types), m.toGenericString()));
}
arguments = new CodeGeneratorArgument[parameterNames.length];
- for (int i = 0 ; i < parameterNames.length; i++) {
+ for (int i = 0; i < parameterNames.length; i++) {
arguments[i] = new CodeGeneratorArgument(parameterNames[i], types[i]);
}
exs = m.getExceptionTypes();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java b/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java
index e94d9f859..a8b5f4b4c 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java
@@ -21,10 +21,10 @@ import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.Path;
public enum DotDrillType {
- VIEW
+ VIEW;
// ,FORMAT
// ,STATS
- ;
+
private final String ending;
@@ -70,4 +70,4 @@ public enum DotDrillType {
b.append("}.drill");
DOT_DRILL_GLOB = b.toString();
}
-} \ No newline at end of file
+}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java
index 4486972ee..373cdf69a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java
@@ -778,7 +778,7 @@ public class EvaluationVisitor {
setBlock.assign(out.getValue(), JExpr.lit(1));
} else {
assert (e == null);
- eval.assign(out.getValue(), JExpr.lit(1)) ;
+ eval.assign(out.getValue(), JExpr.lit(1));
}
generator.unNestEvalBlock(); // exit from nested block
@@ -841,7 +841,7 @@ public class EvaluationVisitor {
setBlock.assign(out.getValue(), JExpr.lit(0));
} else {
assert (e == null);
- eval.assign(out.getValue(), JExpr.lit(0)) ;
+ eval.assign(out.getValue(), JExpr.lit(0));
}
generator.unNestEvalBlock(); // exit from nested block.
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java
index fdc469724..6263dec52 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java
@@ -42,7 +42,7 @@ public class SizedJBlock {
}
public void incCounter() {
- this.count ++;
+ this.count++;
}
public int getCount() {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java
index 1a5df670f..59638922a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java
@@ -93,13 +93,13 @@ class DrillAggFuncHolder extends DrillFuncHolder {
//Loop through all workspace vectors, to get the minimum of size of all workspace vectors.
JVar sizeVar = setupBlock.decl(g.getModel().INT, "vectorSize", JExpr.lit(Integer.MAX_VALUE));
JClass mathClass = g.getModel().ref(Math.class);
- for (int id = 0; id < getWorkspaceVars().length; id ++) {
+ for (int id = 0; id < getWorkspaceVars().length; id++) {
if (!getWorkspaceVars()[id].isInject()) {
setupBlock.assign(sizeVar,mathClass.staticInvoke("min").arg(sizeVar).arg(g.getWorkspaceVectors().get(getWorkspaceVars()[id]).invoke("getValueCapacity")));
}
}
- for(int i =0 ; i < getWorkspaceVars().length; i++) {
+ for (int i = 0; i < getWorkspaceVars().length; i++) {
if (!getWorkspaceVars()[i].isInject()) {
setupBlock.assign(workspaceJVars[i], JExpr._new(g.getHolderType(getWorkspaceVars()[i].getMajorType())));
}
@@ -156,7 +156,7 @@ class DrillAggFuncHolder extends DrillFuncHolder {
private JVar[] declareWorkspaceVectors(ClassGenerator<?> g) {
JVar[] workspaceJVars = new JVar[getWorkspaceVars().length];
- for(int i =0 ; i < getWorkspaceVars().length; i++){
+ for (int i = 0; i < getWorkspaceVars().length; i++) {
if (getWorkspaceVars()[i].isInject()) {
workspaceJVars[i] = g.declareClassField("work", g.getModel()._ref(getWorkspaceVars()[i].getType()));
g.getBlock(BlockType.SETUP).assign(workspaceJVars[i], g.getMappingSet().getIncoming().invoke("getContext").invoke("getManagedBuffer"));
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java
index 412674f13..4789a48d4 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java
@@ -346,7 +346,7 @@ public class Hash32Functions {
}
}
- @FunctionTemplate(names = {"hash", "hash32" ,"hash32AsDouble"}, scope = FunctionScope.SIMPLE, nulls = FunctionTemplate.NullHandling.INTERNAL)
+ @FunctionTemplate(names = {"hash", "hash32", "hash32AsDouble"}, scope = FunctionScope.SIMPLE, nulls = FunctionTemplate.NullHandling.INTERNAL)
public static class NullableTimeHash implements DrillSimpleFunc {
@Param NullableTimeHolder in;
@Output IntHolder out;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java
index 43098486c..c23a1a298 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java
@@ -498,7 +498,7 @@ public class StringFunctions{
for (int id = input.start; id < input.end; id++) {
byte currentByte = input.buffer.getByte(id);
- out.buffer.setByte(id - input.start, Character.toLowerCase(currentByte)) ;
+ out.buffer.setByte(id - input.start, Character.toLowerCase(currentByte));
}
}
}
@@ -529,7 +529,7 @@ public class StringFunctions{
for (int id = input.start; id < input.end; id++) {
byte currentByte = input.buffer.getByte(id);
- out.buffer.setByte(id - input.start, Character.toUpperCase(currentByte)) ;
+ out.buffer.setByte(id - input.start, Character.toUpperCase(currentByte));
}
}
}
@@ -711,9 +711,9 @@ public class StringFunctions{
final int charCount = org.apache.drill.exec.expr.fn.impl.StringFunctionUtil.getUTF8CharLength(string.buffer, string.start, string.end);
final int charLen;
if (length.value > 0) {
- charLen = Math.min((int)length.value, charCount); //left('abc', 5) -> 'abc'
+ charLen = Math.min((int) length.value, charCount); //left('abc', 5) -> 'abc'
} else if (length.value < 0) {
- charLen = Math.max(0, charCount + (int)length.value) ; // left('abc', -5) ==> ''
+ charLen = Math.max(0, charCount + (int) length.value); // left('abc', -5) ==> ''
} else {
charLen = 0;
}
@@ -830,7 +830,7 @@ public class StringFunctions{
if (j == from.end ) {
//find a true match ("from" is not empty), copy entire "to" string to out buffer
- for (int k = to.start ; k < to.end; k++) {
+ for (int k = to.start; k < to.end; k++) {
out.buffer.setByte(out.end++, to.buffer.getByte(k));
}
@@ -921,7 +921,7 @@ public class StringFunctions{
(currentByte & 0xE0) == 0xC0 || // 2-byte char. First byte is 110xxxxx
(currentByte & 0xF0) == 0xE0 || // 3-byte char. First byte is 1110xxxx
(currentByte & 0xF8) == 0xF0) { //4-byte char. First byte is 11110xxx
- count ++; //Advance the counter, since we find one char.
+ count++; //Advance the counter, since we find one char.
}
out.buffer.setByte(out.end++, currentByte);
}
@@ -1070,7 +1070,7 @@ public class StringFunctions{
(currentByte & 0xE0) == 0xC0 || // 2-byte char. First byte is 110xxxxx
(currentByte & 0xF0) == 0xE0 || // 3-byte char. First byte is 1110xxxx
(currentByte & 0xF8) == 0xF0) { //4-byte char. First byte is 11110xxx
- count ++; //Advance the counter, since we find one char.
+ count++; //Advance the counter, since we find one char.
}
out.buffer.setByte(out.end++, currentByte);
}
@@ -1429,7 +1429,7 @@ public class StringFunctions{
@Override
public void eval() {
- out.buffer = buffer = buffer.reallocIfNeeded( (left.end - left.start) + (right.end - right.start));
+ out.buffer = buffer = buffer.reallocIfNeeded((left.end - left.start) + (right.end - right.start));
out.start = out.end = 0;
int id = 0;
@@ -1493,7 +1493,7 @@ public class StringFunctions{
@Override
public void eval() {
- out.buffer = buffer = buffer.reallocIfNeeded( (left.end - left.start) + (right.end - right.start));;
+ out.buffer = buffer = buffer.reallocIfNeeded((left.end - left.start) + (right.end - right.start));
out.start = out.end = 0;
int id = 0;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java
index e38489236..1bb2512de 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java
@@ -44,7 +44,8 @@ public class JsonConvertTo {
private JsonConvertTo(){}
- @FunctionTemplate(names = { "convert_toJSON", "convert_toSIMPLEJSON" } , scope = FunctionScope.SIMPLE, nulls = NullHandling.NULL_IF_NULL,
+ @FunctionTemplate(names = { "convert_toJSON", "convert_toSIMPLEJSON" },
+ scope = FunctionScope.SIMPLE, nulls = NullHandling.NULL_IF_NULL,
outputWidthCalculatorType = FunctionTemplate.OutputWidthCalculatorType.CUSTOM_FIXED_WIDTH_DEFUALT)
public static class ConvertToJson implements DrillSimpleFunc{
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java
index e127391fe..c07e6ed37 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java
@@ -316,7 +316,7 @@ public class FunctionRegistryHolder {
final String functionName = function.getName();
Queue<String> jarFunctions = jar.get(functionName);
if (jarFunctions == null) {
- jarFunctions = Queues.newConcurrentLinkedQueue();;
+ jarFunctions = Queues.newConcurrentLinkedQueue();
jar.put(functionName, jarFunctions);
}
final String functionSignature = function.getSignature();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java
index 51f34a017..da988dec4 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java
@@ -104,6 +104,6 @@ public class HashAggregate extends AbstractSingle {
public boolean isBufferedOperator(QueryContext queryContext) {
// In case forced to use a single partition - do not consider this a buffered op (when memory is divided)
return queryContext == null ||
- 1 < (int)queryContext.getOptions().getOption(ExecConstants.HASHAGG_NUM_PARTITIONS_VALIDATOR) ;
+ 1 < (int) queryContext.getOptions().getOption(ExecConstants.HASHAGG_NUM_PARTITIONS_VALIDATOR);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java
index 48d977ea1..146bb7fd6 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java
@@ -88,6 +88,6 @@ public class HashJoinPOP extends AbstractJoinPop {
public boolean isBufferedOperator(QueryContext queryContext) {
// In case forced to use a single partition - do not consider this a buffered op (when memory is divided)
return queryContext == null ||
- 1 < (int)queryContext.getOptions().getOption(ExecConstants.HASHJOIN_NUM_PARTITIONS_VALIDATOR) ;
+ 1 < (int) queryContext.getOptions().getOption(ExecConstants.HASHJOIN_NUM_PARTITIONS_VALIDATOR);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java
index 4bbfa05a1..1954c79a7 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java
@@ -154,7 +154,13 @@ public abstract class HashAggTemplate implements HashAggregator {
private int cycleNum = 0; // primary, secondary, tertiary, etc.
private int originalPartition = -1; // the partition a secondary reads from
- private static class SpilledPartition { public int spilledBatches; public String spillFile; int cycleNum; int origPartn; int prevOrigPartn; }
+ private static class SpilledPartition {
+ public int spilledBatches;
+ public String spillFile;
+ int cycleNum;
+ int origPartn;
+ int prevOrigPartn;
+ }
private ArrayList<SpilledPartition> spilledPartitionsList;
private int operatorId; // for the spill file name
@@ -193,7 +199,6 @@ public abstract class HashAggTemplate implements HashAggregator {
AVG_OUTPUT_BATCH_BYTES,
AVG_OUTPUT_ROW_BYTES,
OUTPUT_RECORD_COUNT;
- ;
@Override
public int metricId() {
@@ -333,7 +338,7 @@ public abstract class HashAggTemplate implements HashAggregator {
is2ndPhase = hashAggrConfig.getAggPhase() == AggPrelBase.OperatorPhase.PHASE_2of2;
isTwoPhase = hashAggrConfig.getAggPhase() != AggPrelBase.OperatorPhase.PHASE_1of1;
- is1stPhase = isTwoPhase && ! is2ndPhase ;
+ is1stPhase = isTwoPhase && !is2ndPhase;
canSpill = isTwoPhase; // single phase can not spill
// Typically for testing - force a spill after a partition has more than so many batches
@@ -410,8 +415,8 @@ public abstract class HashAggTemplate implements HashAggregator {
updateEstMaxBatchSize(incoming);
}
// create "reserved memory" and adjust the memory limit down
- reserveValueBatchMemory = reserveOutgoingMemory = estValuesBatchSize ;
- long newMemoryLimit = allocator.getLimit() - reserveValueBatchMemory - reserveOutgoingMemory ;
+ reserveValueBatchMemory = reserveOutgoingMemory = estValuesBatchSize;
+ long newMemoryLimit = allocator.getLimit() - reserveValueBatchMemory - reserveOutgoingMemory;
long memAvail = newMemoryLimit - allocator.getAllocatedMemory();
if ( memAvail <= 0 ) { throw new OutOfMemoryException("Too little memory available"); }
allocator.setLimit(newMemoryLimit);
@@ -458,9 +463,9 @@ public abstract class HashAggTemplate implements HashAggregator {
bitsInMask = Integer.bitCount(partitionMask); // e.g. 0x1F -> 5
// Create arrays (one entry per partition)
- htables = new HashTable[numPartitions] ;
- batchHolders = (ArrayList<BatchHolder>[]) new ArrayList<?>[numPartitions] ;
- outBatchIndex = new int[numPartitions] ;
+ htables = new HashTable[numPartitions];
+ batchHolders = (ArrayList<BatchHolder>[]) new ArrayList<?>[numPartitions];
+ outBatchIndex = new int[numPartitions];
writers = new Writer[numPartitions];
spilledBatchesCount = new int[numPartitions];
spillFiles = new String[numPartitions];
@@ -486,7 +491,11 @@ public abstract class HashAggTemplate implements HashAggregator {
this.batchHolders[i] = new ArrayList<BatchHolder>(); // First BatchHolder is created when the first put request is received.
}
// Initialize the value vectors in the generated code (which point to the incoming or outgoing fields)
- try { htables[0].updateBatches(); } catch (SchemaChangeException sc) { throw new UnsupportedOperationException(sc); };
+ try {
+ htables[0].updateBatches();
+ } catch (SchemaChangeException sc) {
+ throw new UnsupportedOperationException(sc);
+ }
}
/**
* get new incoming: (when reading spilled files like an "incoming")
@@ -689,7 +698,7 @@ public abstract class HashAggTemplate implements HashAggregator {
// Either flag buildComplete or handleEmit (or earlyOutput) would cause returning of
// the outgoing batch downstream (see innerNext() in HashAggBatch).
- buildComplete = true ; // now should go and return outgoing
+ buildComplete = true; // now should go and return outgoing
if ( handleEmit ) {
buildComplete = false; // This was not a real NONE - more incoming is expected
@@ -939,7 +948,7 @@ public abstract class HashAggTemplate implements HashAggregator {
}
// Give the current (if already spilled) some priority
if ( ! tryAvoidCurr && isSpilled(currPart) && ( currPartSize + 1 >= maxSizeSpilled )) {
- maxSizeSpilled = currPartSize ;
+ maxSizeSpilled = currPartSize;
indexMaxSpilled = currPart;
}
// now find the largest non-spilled partition
@@ -948,7 +957,7 @@ public abstract class HashAggTemplate implements HashAggregator {
// Use the largest spilled (if found) as a base line, with a factor of 4
if ( indexMaxSpilled > -1 && maxSizeSpilled > 1 ) {
indexMax = indexMaxSpilled;
- maxSize = 4 * maxSizeSpilled ;
+ maxSize = 4 * maxSizeSpilled;
}
for ( int insp = 0; insp < numPartitions; insp++) {
if ( ! isSpilled(insp) && maxSize < batchHolders[insp].size() ) {
@@ -1159,7 +1168,11 @@ public abstract class HashAggTemplate implements HashAggregator {
originalPartition = sp.origPartn; // used for the filename
logger.trace("Reading back spilled original partition {} as an incoming",originalPartition);
// Initialize .... new incoming, new set of partitions
- try { initializeSetup(newIncoming); } catch (Exception e) { throw new RuntimeException(e); }
+ try {
+ initializeSetup(newIncoming);
+ } catch (Exception e) {
+ throw new RuntimeException(e);
+ }
// update the cycle num if needed
// The current cycle num should always be one larger than in the spilled partition
if ( cycleNum == sp.cycleNum ) {
@@ -1179,7 +1192,7 @@ public abstract class HashAggTemplate implements HashAggregator {
return AggIterOutcome.AGG_RESTART;
}
- partitionToReturn = nextPartitionToReturn ;
+ partitionToReturn = nextPartitionToReturn;
}
@@ -1187,7 +1200,7 @@ public abstract class HashAggTemplate implements HashAggregator {
int numPendingOutput = currPartition.get(currOutBatchIndex).getNumPendingOutput();
// The following accounting is for logging, metrics, etc.
- rowsInPartition += numPendingOutput ;
+ rowsInPartition += numPendingOutput;
if ( ! handlingSpills ) { rowsNotSpilled += numPendingOutput; }
else { rowsSpilledReturned += numPendingOutput; }
if ( earlyOutput ) { rowsReturnedEarly += numPendingOutput; }
@@ -1238,7 +1251,7 @@ public abstract class HashAggTemplate implements HashAggregator {
logger.debug("HASH AGG: Finished (early) re-init partition {}, mem allocated: {}", earlyPartition, allocator.getAllocatedMemory());
}
outBatchIndex[earlyPartition] = 0; // reset, for next time
- earlyOutput = false ; // done with early output
+ earlyOutput = false; // done with early output
}
else if ( handleEmit ) {
// When returning the last outgoing batch (following an incoming EMIT), then replace OK with EMIT
@@ -1290,9 +1303,9 @@ public abstract class HashAggTemplate implements HashAggregator {
*/
private String getOOMErrorMsg(String prefix) {
String errmsg;
- if ( !isTwoPhase ) {
- errmsg = "Single Phase Hash Aggregate operator can not spill." ;
- } else if ( ! canSpill ) { // 2nd phase, with only 1 partition
+ if (!isTwoPhase) {
+ errmsg = "Single Phase Hash Aggregate operator can not spill.";
+ } else if (!canSpill) { // 2nd phase, with only 1 partition
errmsg = "Too little memory available to operator to facilitate spilling.";
} else { // a bug ?
errmsg = prefix + " OOM at " + (is2ndPhase ? "Second Phase" : "First Phase") + ". Partitions: " + numPartitions +
@@ -1353,9 +1366,11 @@ public abstract class HashAggTemplate implements HashAggregator {
}
// right shift hash code for secondary (or tertiary...) spilling
- for (int i = 0; i < cycleNum; i++) { hashCode >>>= bitsInMask; }
+ for (int i = 0; i < cycleNum; i++) {
+ hashCode >>>= bitsInMask;
+ }
- int currentPartition = hashCode & partitionMask ;
+ int currentPartition = hashCode & partitionMask;
hashCode >>>= bitsInMask;
HashTable.PutStatus putStatus = null;
long allocatedBeforeHTput = allocator.getAllocatedMemory();
@@ -1398,7 +1413,7 @@ public abstract class HashAggTemplate implements HashAggregator {
throw new UnsupportedOperationException("Unexpected schema change", e);
}
long allocatedBeforeAggCol = allocator.getAllocatedMemory();
- boolean needToCheckIfSpillIsNeeded = allocatedBeforeAggCol > allocatedBeforeHTput ;
+ boolean needToCheckIfSpillIsNeeded = allocatedBeforeAggCol > allocatedBeforeHTput;
// Add an Aggr batch if needed:
//
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java
index f58be8929..4c54650cf 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java
@@ -44,7 +44,7 @@ public interface HashAggregator {
// For returning results from outputCurrentBatch
// OK - batch returned, NONE - end of data, RESTART - call again, EMIT - like OK but EMIT
- enum AggIterOutcome { AGG_OK, AGG_NONE, AGG_RESTART , AGG_EMIT }
+ enum AggIterOutcome { AGG_OK, AGG_NONE, AGG_RESTART, AGG_EMIT }
void setup(HashAggregate hashAggrConfig, HashTableConfig htConfig, FragmentContext context, OperatorContext oContext, RecordBatch incoming, HashAggBatch outgoing,
LogicalExpression[] valueExprs, List<TypedFieldId> valueFieldIds, TypedFieldId[] keyFieldIds, VectorContainer outContainer, int extraRowBytes) throws SchemaChangeException, IOException, ClassTransformationException;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java
index c78e2c01f..7ebce2b4a 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java
@@ -135,7 +135,7 @@ public class SpilledRecordbatch implements CloseableRecordBatch {
if ( spillStream == null ) {
throw new IllegalStateException("Spill stream was null");
- };
+ }
if ( spillSet.getPosition(spillStream) < 0 ) {
HashAggTemplate.logger.warn("Position is {} for stream {}", spillSet.getPosition(spillStream), spillStream.toString());
@@ -155,7 +155,7 @@ public class SpilledRecordbatch implements CloseableRecordBatch {
throw UserException.dataReadError(e).addContext("Failed reading from a spill file").build(HashAggTemplate.logger);
}
- spilledBatches-- ; // one less batch to read
+ spilledBatches--; // one less batch to read
return IterOutcome.OK;
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java
index 2a64b9307..23fdcc1d2 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java
@@ -49,8 +49,7 @@ public interface StreamingAggregator {
RETURN_OUTCOME,
CLEANUP_AND_RETURN,
UPDATE_AGGREGATOR,
- RETURN_AND_RESET
- ;
+ RETURN_AND_RESET;
}
public abstract void setup(OperatorContext context, RecordBatch incoming, StreamingAggBatch outgoing) throws SchemaChangeException;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java
index 8f0b7abcb..8c93f1bb6 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java
@@ -28,10 +28,10 @@ public class HashTableStats {
}
public void addStats (HashTableStats newStats) {
- this.numBuckets += newStats.numBuckets ;
- this.numEntries += newStats.numEntries ;
- this.numResizing += newStats.numResizing ;
- this.resizingTime += newStats.resizingTime ;
+ this.numBuckets += newStats.numBuckets;
+ this.numEntries += newStats.numEntries;
+ this.numResizing += newStats.numResizing;
+ this.resizingTime += newStats.resizingTime;
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java
index 83b72d7c7..3c418b99c 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java
@@ -476,7 +476,7 @@ public abstract class HashTableTemplate implements HashTable {
if (tableSize > MAXIMUM_CAPACITY) {
tableSize = MAXIMUM_CAPACITY;
}
- originalTableSize = tableSize ; // retain original size
+ originalTableSize = tableSize; // retain original size
threshold = (int) Math.ceil(tableSize * loadf);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java
index cd58bfd5d..365451c02 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java
@@ -113,7 +113,7 @@ public abstract class FlattenTemplate implements Flattener {
} catch (OversizedAllocationException ex) {
// unable to flatten due to a soft buffer overflow. split the batch here and resume execution.
logger.debug("Reached allocation limit. Splitting the batch at input index: {} - inner index: {} - current completed index: {}",
- valueIndexLocal, innerValueIndexLocal, currentInnerValueIndexLocal) ;
+ valueIndexLocal, innerValueIndexLocal, currentInnerValueIndexLocal);
/*
* TODO
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java
index d4d4f927e..b1ea96f05 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java
@@ -776,7 +776,7 @@ public class HashJoinBatch extends AbstractBinaryRecordBatch<HashJoinPOP> {
for (int ind = 0; ind < currentRecordCount; ind++) {
int hashCode = ( cycleNum == 0 ) ? partitions[0].getBuildHashCode(ind)
: read_right_HV_vector.getAccessor().get(ind); // get the hash value from the HV column
- int currPart = hashCode & partitionMask ;
+ int currPart = hashCode & partitionMask;
hashCode >>>= bitsInMask;
// Append the new inner row to the appropriate partition; spill (that partition) if needed
partitions[currPart].appendInnerRow(buildBatch.getContainer(), ind, hashCode, buildCalc); // may spill if needed
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java
index 46f2fa369..639f757ec 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java
@@ -162,7 +162,7 @@ public abstract class HashJoinProbeTemplate implements HashJoinProbe {
*/
private int appendBuild(VectorContainer buildSrcContainer, int buildSrcIndex) {
// "- 1" to skip the last "hash values" added column
- int lastColIndex = buildSrcContainer.getNumberOfColumns() - 1 ;
+ int lastColIndex = buildSrcContainer.getNumberOfColumns() - 1;
for (int vectorIndex = 0; vectorIndex < lastColIndex; vectorIndex++) {
ValueVector destVector = container.getValueVector(vectorIndex).getValueVector();
ValueVector srcVector = buildSrcContainer.getValueVector(vectorIndex).getValueVector();
@@ -292,7 +292,7 @@ public abstract class HashJoinProbeTemplate implements HashJoinProbe {
int hashCode = ( cycleNum == 0 ) ?
partitions[0].getProbeHashCode(recordsProcessed)
: read_left_HV_vector.getAccessor().get(recordsProcessed);
- int currBuildPart = hashCode & partitionMask ;
+ int currBuildPart = hashCode & partitionMask;
hashCode >>>= bitsInMask;
// Set and keep the current partition (may be used again on subsequent probe calls as
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java
index b9745371e..52871e2b8 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java
@@ -235,7 +235,7 @@ public class JoinUtils {
if (currentrel instanceof DrillAggregateRel) {
agg = (DrillAggregateRel)currentrel;
} else if (currentrel instanceof RelSubset) {
- currentrel = ((RelSubset) currentrel).getBest() ;
+ currentrel = ((RelSubset) currentrel).getBest();
} else if (currentrel instanceof DrillLimitRel) {
// TODO: Improve this check when DRILL-5691 is fixed.
// The problem is that RelMdMaxRowCount currently cannot be used
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java
index ff33e2f3c..0eb15a7d0 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java
@@ -626,8 +626,8 @@ public class LateralJoinBatch extends AbstractBinaryRecordBatch<LateralJoinPOP>
if (leftUpstream == EMIT || leftUpstream == OK_NEW_SCHEMA) {
break;
} else {
- logger.debug("Output batch still has some space left, getting new batches from left and right. OutIndex: {}"
- , outputIndex);
+ logger.debug("Output batch still has some space left, getting new batches from left and right. OutIndex: {}",
+ outputIndex);
// Get both left batch and the right batch and make sure indexes are properly set
leftUpstream = processLeftBatch();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java
index 034d6c23d..5049b067e 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java
@@ -269,7 +269,7 @@ public class PartitionSenderRootExec extends BaseRootExec {
// set up partitioning function
final LogicalExpression expr = operator.getExpr();
final ErrorCollector collector = new ErrorCollectorImpl();
- final ClassGenerator<Partitioner> cg ;
+ final ClassGenerator<Partitioner> cg;
cg = CodeGenerator.getRoot(Partitioner.TEMPLATE_DEFINITION, context.getOptions());
cg.getCodeGenerator().plainJavaCapable(true);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java
index 02d2f1833..f0816b625 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java
@@ -54,7 +54,7 @@ public class UnnestImpl implements Unnest {
private RepeatedValueVector.RepeatedAccessor accessor;
private RecordBatch outgoing;
- private IntVector rowIdVector ; // Allocated and owned by the UnnestRecordBatch
+ private IntVector rowIdVector; // Allocated and owned by the UnnestRecordBatch
private IntVector.Mutator rowIdVectorMutator;
/**
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java
index 68b546b95..5d2721a81 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java
@@ -266,7 +266,7 @@ public class SortMemoryManager {
memoryLimit = (configMemoryLimit == 0) ? opMemoryLimit
: Math.min(opMemoryLimit, configMemoryLimit);
- preferredSpillBatchSize = config.spillBatchSize();;
+ preferredSpillBatchSize = config.spillBatchSize();
preferredMergeBatchSize = config.mergeBatchSize();
// Initialize the buffer memory limit for the first batch.
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java
index acc9556f8..24e270a4f 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java
@@ -347,8 +347,7 @@ public abstract class ColumnState {
.attribute("addVersion", addVersion)
.attribute("state", state)
.attributeIdentity("writer", writer)
- .attribute("vectorState")
- ;
+ .attribute("vectorState");
vectorState.dump(format);
format.endObject();
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java
index c7c6fdc5d..2783c5b30 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java
@@ -795,8 +795,7 @@ public class ResultSetLoaderImpl implements ResultSetLoader, LoaderInternals {
.attribute("activeSchemaVersion", activeSchemaVersion)
.attribute("harvestSchemaVersion", harvestSchemaVersion)
.attribute("pendingRowCount", pendingRowCount)
- .attribute("targetRowCount", targetRowCount)
- ;
+ .attribute("targetRowCount", targetRowCount);
format.attribute("root");
rootState.dump(format);
format.attribute("rootWriter");
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java
index b70c764e2..cf0125692 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java
@@ -44,7 +44,7 @@ public abstract class AbstractPartitionDescriptor implements PartitionDescriptor
* Create sublists of the partition locations, each sublist of size
* at most {@link PartitionDescriptor#PARTITION_BATCH_SIZE}
*/
- protected abstract void createPartitionSublists() ;
+ protected abstract void createPartitionSublists();
/**
* Iterator that traverses over the super list of partition locations and
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java
index 216c8d27b..4cd9ed3c5 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java
@@ -65,7 +65,7 @@ public class StarColumnHelper {
}
public static boolean isPrefixedStarColumn(String fieldName) {
- return fieldName.indexOf(PREFIXED_STAR_COLUMN) > 0 ; // the delimiter * starts at none-zero position.
+ return fieldName.indexOf(PREFIXED_STAR_COLUMN) > 0; // the delimiter * starts at none-zero position.
}
public static boolean isNonPrefixedStarColumn(String fieldName) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java
index 10c4738ba..862fb59fd 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java
@@ -137,7 +137,7 @@ public abstract class DrillJoinRelBase extends Join implements DrillRelNode {
// just to make sure Cartesian Join is more expensive
// than Non-Cartesian Join.
- final int keySize = 1 ; // assume having 1 join key, when estimate join cost.
+ final int keySize = 1; // assume having 1 join key, when estimate join cost.
final DrillCostBase cost = (DrillCostBase) computeHashJoinCostWithKeySize(planner, keySize, mq).multiplyBy(mulFactor);
// Cartesian join row count will be product of two inputs. The other factors come from the above estimated DrillCost.
@@ -197,7 +197,7 @@ public abstract class DrillJoinRelBase extends Join implements DrillRelNode {
) * buildRowCount * factor;
double cpuCost = joinConditionCost * (probeRowCount) // probe size determine the join condition comparison cost
- + cpuCostBuild + cpuCostProbe ;
+ + cpuCostBuild + cpuCostProbe;
DrillCostFactory costFactory = (DrillCostFactory) planner.getCostFactory();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java
index 7d070b6ac..fde3896c1 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java
@@ -84,7 +84,7 @@ public abstract class DrillLimitRelBase extends SingleRel implements DrillRelNod
@Override
public double estimateRowCount(RelMetadataQuery mq) {
- int off = offset != null ? RexLiteral.intValue(offset) : 0 ;
+ int off = offset != null? RexLiteral.intValue(offset): 0;
if (fetch == null) {
// If estimated rowcount is less than offset return 0
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java
index b7881c2c9..6c1b14330 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java
@@ -60,7 +60,7 @@ import com.google.common.collect.Lists;
* Base class for logical and physical Project implemented in Drill
*/
public abstract class DrillProjectRelBase extends Project implements DrillRelNode {
- private final int nonSimpleFieldCount ;
+ private final int nonSimpleFieldCount;
protected DrillProjectRelBase(Convention convention, RelOptCluster cluster, RelTraitSet traits, RelNode child, List<? extends RexNode> exps,
RelDataType rowType) {
@@ -128,10 +128,10 @@ public abstract class DrillProjectRelBase extends Project implements DrillRelNod
for (RexNode expr : this.getProjects()) {
if (expr instanceof RexInputRef) {
// Simple Field reference.
- cnt ++;
+ cnt++;
} else if (expr instanceof RexCall && expr.accept(complexFieldIdentifer)) {
// Complex field with named segments only.
- cnt ++;
+ cnt++;
}
}
return cnt;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java
index ba55faeb7..f64cd7708 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java
@@ -169,7 +169,7 @@ public class DrillCostBase implements DrillRelOptCost {
|| (this.io == Double.POSITIVE_INFINITY)
|| (this.network == Double.POSITIVE_INFINITY)
|| (this.rowCount == Double.POSITIVE_INFINITY)
- || (this.memory == Double.POSITIVE_INFINITY) ;
+ || (this.memory == Double.POSITIVE_INFINITY);
}
@Override
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java
index c96f4e6eb..7497783e3 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java
@@ -51,7 +51,7 @@ public class DrillFilterRel extends DrillFilterRelBase implements DrillRel {
}
public static DrillFilterRel create(RelNode child, RexNode condition) {
- return new DrillFilterRel(child.getCluster(), child.getTraitSet(), child, condition) ;
+ return new DrillFilterRel(child.getCluster(), child.getTraitSet(), child, condition);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java
index 2cd60688c..5f2b02c29 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java
@@ -301,10 +301,15 @@ public class DrillOptiq {
case "CHAR":
castType = Types.required(MinorType.VARCHAR).toBuilder().setPrecision(call.getType().getPrecision()).build();
break;
-
- case "INTEGER": castType = Types.required(MinorType.INT); break;
- case "FLOAT": castType = Types.required(MinorType.FLOAT4); break;
- case "DOUBLE": castType = Types.required(MinorType.FLOAT8); break;
+ case "INTEGER":
+ castType = Types.required(MinorType.INT);
+ break;
+ case "FLOAT":
+ castType = Types.required(MinorType.FLOAT4);
+ break;
+ case "DOUBLE":
+ castType = Types.required(MinorType.FLOAT8);
+ break;
case "DECIMAL":
if (!context.getPlannerSettings().getOptions().getOption(PlannerSettings.ENABLE_DECIMAL_DATA_TYPE_KEY).bool_val) {
throw UserException
@@ -327,7 +332,9 @@ public class DrillOptiq {
case "INTERVAL_YEAR":
case "INTERVAL_YEAR_MONTH":
- case "INTERVAL_MONTH": castType = Types.required(MinorType.INTERVALYEAR); break;
+ case "INTERVAL_MONTH":
+ castType = Types.required(MinorType.INTERVALYEAR);
+ break;
case "INTERVAL_DAY":
case "INTERVAL_DAY_HOUR":
case "INTERVAL_DAY_MINUTE":
@@ -337,11 +344,19 @@ public class DrillOptiq {
case "INTERVAL_HOUR_SECOND":
case "INTERVAL_MINUTE":
case "INTERVAL_MINUTE_SECOND":
- case "INTERVAL_SECOND": castType = Types.required(MinorType.INTERVALDAY); break;
- case "BOOLEAN": castType = Types.required(MinorType.BIT); break;
- case "BINARY": castType = Types.required(MinorType.VARBINARY); break;
- case "ANY": return arg; // Type will be same as argument.
- default: castType = Types.required(MinorType.valueOf(call.getType().getSqlTypeName().getName()));
+ case "INTERVAL_SECOND":
+ castType = Types.required(MinorType.INTERVALDAY);
+ break;
+ case "BOOLEAN":
+ castType = Types.required(MinorType.BIT);
+ break;
+ case "BINARY":
+ castType = Types.required(MinorType.VARBINARY);
+ break;
+ case "ANY":
+ return arg; // Type will be same as argument.
+ default:
+ castType = Types.required(MinorType.valueOf(call.getType().getSqlTypeName().getName()));
}
return FunctionCallFactory.createCast(castType, ExpressionPosition.UNKNOWN, arg);
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java
index 293749ba6..69e9452d7 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java
@@ -61,7 +61,7 @@ public class DrillUnionAllRule extends RelOptRule {
call.transformTo(new DrillUnionRel(union.getCluster(), traits, convertedInputs, union.all,
true /* check compatibility */));
} catch (InvalidRelException e) {
- tracer.warn(e.toString()) ;
+ tracer.warn(e.toString());
}
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java
index c94d2b5a9..a5a6e0330 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java
@@ -50,7 +50,7 @@ public class DrillUnionRel extends DrillUnionRelBase implements DrillRel {
return new DrillUnionRel(getCluster(), traitSet, inputs, all,
false /* don't check compatibility during copy */);
} catch (InvalidRelException e) {
- throw new AssertionError(e) ;
+ throw new AssertionError(e);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java
index ca68a7d1a..ff640a7b6 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java
@@ -49,7 +49,7 @@ public abstract class AggPrelBase extends DrillAggregateRelBase implements Prel
public enum OperatorPhase {PHASE_1of1, PHASE_1of2, PHASE_2of2}
- protected OperatorPhase operPhase = OperatorPhase.PHASE_1of1 ; // default phase
+ protected OperatorPhase operPhase = OperatorPhase.PHASE_1of1; // default phase
protected List<NamedExpression> keys = Lists.newArrayList();
protected List<NamedExpression> aggExprs = Lists.newArrayList();
protected List<AggregateCall> phase2AggCallList = Lists.newArrayList();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java
index 2c043689b..caec426be 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java
@@ -56,7 +56,7 @@ public class BroadcastExchangePrel extends ExchangePrel{
final double inputRows = mq.getRowCount(child);
final int rowWidth = child.getRowType().getFieldCount() * DrillCostBase.AVG_FIELD_WIDTH;
- final double cpuCost = broadcastFactor * DrillCostBase.SVR_CPU_COST * inputRows ;
+ final double cpuCost = broadcastFactor * DrillCostBase.SVR_CPU_COST * inputRows;
final double networkCost = broadcastFactor * DrillCostBase.BYTE_NETWORK_COST * inputRows * rowWidth * numEndPoints;
return new DrillCostBase(inputRows, cpuCost, 0, networkCost);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java
index d10021cc7..b250d5a94 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java
@@ -96,7 +96,7 @@ public class DrillDistributionTrait implements RelTrait {
@Override
public int hashCode() {
- return fields == null ? type.hashCode() : type.hashCode() | fields.hashCode() << 4 ;
+ return fields == null? type.hashCode(): type.hashCode() | fields.hashCode() << 4;
}
@Override
@@ -106,7 +106,7 @@ public class DrillDistributionTrait implements RelTrait {
}
if (obj instanceof DrillDistributionTrait) {
DrillDistributionTrait that = (DrillDistributionTrait) obj;
- return this.type == that.type && this.fields.equals(that.fields) ;
+ return this.type == that.type && this.fields.equals(that.fields);
}
return false;
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java
index 19499d67b..ade0fb349 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java
@@ -91,7 +91,7 @@ public class HashAggPrule extends AggPruleBase {
createTransformRequest(call, aggregate, input, traits);
if (create2PhasePlan(call, aggregate)) {
- traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ;
+ traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL);
RelNode convertedInput = convert(input, traits);
new TwoPhaseSubset(call, distOnAllKeys).go(aggregate, convertedInput);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java
index 2272a9b97..c9be46f9c 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java
@@ -38,7 +38,7 @@ public class HashToMergeExchangePrel extends ExchangePrel {
private final List<DistributionField> distFields;
private int numEndPoints = 0;
- private final RelCollation collation ;
+ private final RelCollation collation;
public HashToMergeExchangePrel(RelOptCluster cluster, RelTraitSet traitSet, RelNode input,
List<DistributionField> fields,
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java
index a03c2f5d4..223882cc1 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java
@@ -40,7 +40,7 @@ import org.apache.drill.exec.server.options.OptionManager;
public class SingleMergeExchangePrel extends ExchangePrel {
- private final RelCollation collation ;
+ private final RelCollation collation;
public SingleMergeExchangePrel(RelOptCluster cluster, RelTraitSet traitSet, RelNode input, RelCollation collation) {
super(cluster, traitSet, input);
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java
index 85f516a66..99b6cde45 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java
@@ -70,7 +70,7 @@ public class StreamAggPrule extends AggPruleBase {
final RelTraitSet singleDistTrait = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL).plus(singleDist);
if (create2PhasePlan(call, aggregate)) {
- traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ;
+ traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL);
RelNode convertedInput = convert(input, traits);
new SubsetTransformer<DrillAggregateRel, InvalidRelException>(call){
@@ -138,7 +138,7 @@ public class StreamAggPrule extends AggPruleBase {
// createTransformRequest(call, aggregate, input, traits);
if (create2PhasePlan(call, aggregate)) {
- traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ;
+ traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL);
RelNode convertedInput = convert(input, traits);
new SubsetTransformer<DrillAggregateRel, InvalidRelException>(call){
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java
index 43e030072..eddbe4f34 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java
@@ -84,7 +84,9 @@ class NumberingRelWriter implements RelWriter {
s.append(" ");
if (id != null && id.opId == 0) {
- for(int i =0; i < spacer.get(); i++){ s.append('-');}
+ for (int i = 0; i < spacer.get(); i++) {
+ s.append('-');
+ }
}else{
spacer.spaces(s);
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java
index ac491e920..124857823 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java
@@ -248,7 +248,7 @@ public class StarColumnConverter extends BasePrelVisitor<Prel, Void, RuntimeExce
for (String s : names) {
if (uniqueNames.contains(s)) {
- for (int i = 0; ; i++ ) {
+ for (int i = 0;; i++) {
s = s + i;
if (! origNames.contains(s) && ! uniqueNames.contains(s)) {
break;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java b/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java
index 9bb574ce2..90de9af0e 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java
@@ -42,7 +42,7 @@ public interface Controller extends AutoCloseable {
* @param node
* @return
*/
- public ControlTunnel getTunnel(DrillbitEndpoint node) ;
+ public ControlTunnel getTunnel(DrillbitEndpoint node);
public DrillbitEndpoint start(DrillbitEndpoint partialEndpoint, boolean allowPortHunting)
throws DrillbitStartupException;
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java
index bf005b12e..22d0c7b50 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java
@@ -65,7 +65,7 @@ public class ProfileWrapper {
this.profile = profile;
this.id = profile.hasQueryId() ? profile.getQueryId() : QueryIdHelper.getQueryId(profile.getId());
//Generating Operator Name map (DRILL-6140)
- String profileTextPlan = profile.hasPlan() ? profile.getPlan() : "" ;
+ String profileTextPlan = profile.hasPlan()? profile.getPlan(): "";
generateOpMap(profileTextPlan);
final List<FragmentWrapper> fragmentProfiles = new ArrayList<>();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java
index 5a2a37b6c..cddb75ac0 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java
@@ -72,6 +72,6 @@ public class SimpleDurationFormat {
return (days > 0 ? days + " day " : "") +
((hours + days) > 0 ? hours + " hr " : "") +
((minutes + hours + days) > 0 ? String.format("%02d min ", minutes) : "") +
- seconds + "." + String.format("%03d sec", milliSeconds) ;
+ seconds + "." + String.format("%03d sec", milliSeconds);
}
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java b/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java
index 28cbe075b..c3ff0c5db 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java
@@ -253,8 +253,7 @@ public abstract class SSLConfig {
.append("\n\ttrustStorePassword: ").append(getPrintablePassword(getTrustStorePassword()))
.append("\n\thandshakeTimeout: ").append(getHandshakeTimeout())
.append("\n\tdisableHostVerification: ").append(disableHostVerification())
- .append("\n\tdisableCertificateVerification: ").append(disableCertificateVerification())
- ;
+ .append("\n\tdisableCertificateVerification: ").append(disableCertificateVerification());
}
return sb.toString();
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java
index d35743fce..fba80e507 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java
@@ -51,7 +51,7 @@ public interface JsonProcessor {
String msg,
Object... args);
- public boolean ignoreJSONParseError() ;
+ public boolean ignoreJSONParseError();
public void setIgnoreJSONParseErrors(boolean ignoreJSONParseErrors);
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
index ef8f861e5..6d52b6d02 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java
@@ -224,7 +224,7 @@ public class HeaderBuilder extends TextOutput {
// "col", "col_2", "col_2_2", "col_2_2_2".
// No mapping scheme is perfect...
- for (int l = 2; ; l++) {
+ for (int l = 2;; l++) {
final String rewritten = header + "_" + l;
key = rewritten.toLowerCase();
if (! idents.contains(key)) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
index a0043efc3..9ee86ca03 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java
@@ -96,7 +96,7 @@ final class TextInput {
this.lineSeparator = settings.getNewLineDelimiter();
byte normalizedLineSeparator = settings.getNormalizedNewLine();
Preconditions.checkArgument(input instanceof Seekable, "Text input only supports an InputStream that supports Seekable.");
- boolean isCompressed = input instanceof CompressionInputStream ;
+ boolean isCompressed = input instanceof CompressionInputStream;
Preconditions.checkArgument(!isCompressed || startPos == 0, "Cannot use split on compressed stream.");
// splits aren't allowed with compressed data. The split length will be the compressed size which means we'll normally end prematurely.
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java
index c4f89699a..c684e7a52 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java
@@ -130,9 +130,15 @@ public class Records {
break;
// 2. SqlTypeName enumerators whose names (currently) do not match SQL's
// values for DATA_TYPE:
- case CHAR: this.DATA_TYPE = "CHARACTER"; break;
- case VARCHAR: this.DATA_TYPE = "CHARACTER VARYING"; break;
- case VARBINARY: this.DATA_TYPE = "BINARY VARYING"; break;
+ case CHAR:
+ this.DATA_TYPE = "CHARACTER";
+ break;
+ case VARCHAR:
+ this.DATA_TYPE = "CHARACTER VARYING";
+ break;
+ case VARBINARY:
+ this.DATA_TYPE = "BINARY VARYING";
+ break;
case INTERVAL_YEAR:
case INTERVAL_YEAR_MONTH:
case INTERVAL_MONTH:
@@ -145,7 +151,9 @@ public class Records {
case INTERVAL_HOUR_SECOND:
case INTERVAL_MINUTE:
case INTERVAL_MINUTE_SECOND:
- case INTERVAL_SECOND: this.DATA_TYPE = "INTERVAL"; break;
+ case INTERVAL_SECOND:
+ this.DATA_TYPE = "INTERVAL";
+ break;
// 3: SqlTypeName enumerators not yet seen and confirmed or handled.
default:
logger.warn( "Type not handled explicitly (code needs review): "
@@ -212,10 +220,18 @@ public class Records {
this.CHARACTER_OCTET_LENGTH = null;
// This NUMERIC_PRECISION is in bits since NUMERIC_PRECISION_RADIX is 2.
switch ( sqlTypeName ) {
- case TINYINT: NUMERIC_PRECISION = 8; break;
- case SMALLINT: NUMERIC_PRECISION = 16; break;
- case INTEGER: NUMERIC_PRECISION = 32; break;
- case BIGINT: NUMERIC_PRECISION = 64; break;
+ case TINYINT:
+ NUMERIC_PRECISION = 8;
+ break;
+ case SMALLINT:
+ NUMERIC_PRECISION = 16;
+ break;
+ case INTEGER:
+ NUMERIC_PRECISION = 32;
+ break;
+ case BIGINT:
+ NUMERIC_PRECISION = 64;
+ break;
default:
throw new AssertionError(
"Unexpected " + sqlTypeName.getClass().getName() + " value "
@@ -253,9 +269,15 @@ public class Records {
this.CHARACTER_OCTET_LENGTH = null;
// This NUMERIC_PRECISION is in bits since NUMERIC_PRECISION_RADIX is 2.
switch ( sqlTypeName ) {
- case REAL: NUMERIC_PRECISION = 24; break;
- case FLOAT: NUMERIC_PRECISION = 24; break;
- case DOUBLE: NUMERIC_PRECISION = 53; break;
+ case REAL:
+ NUMERIC_PRECISION = 24;
+ break;
+ case FLOAT:
+ NUMERIC_PRECISION = 24;
+ break;
+ case DOUBLE:
+ NUMERIC_PRECISION = 53;
+ break;
default:
throw new AssertionError(
"Unexpected type " + sqlTypeName + " in approximate-types branch" );
@@ -285,7 +307,9 @@ public class Records {
this.INTERVAL_TYPE = null;
this.INTERVAL_PRECISION = null;
switch(sqlTypeName) {
- case DATE: this.COLUMN_SIZE = 10; break;// yyyy-MM-dd
+ case DATE:
+ this.COLUMN_SIZE = 10;
+ break;// yyyy-MM-dd
case TIME: this.COLUMN_SIZE = this.DATETIME_PRECISION == 0
? 8 // HH::mm::ss
: 8 + 1 + this.DATETIME_PRECISION;
@@ -373,8 +397,12 @@ public class Records {
switch(start) {
case YEAR:
switch(end) {
- case YEAR: this.COLUMN_SIZE = INTERVAL_PRECISION + 2; break;// P..Y
- case MONTH: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 5; break; // P..Y12M
+ case YEAR:
+ this.COLUMN_SIZE = INTERVAL_PRECISION + 2;
+ break;// P..Y
+ case MONTH:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 5;
+ break; // P..Y12M
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -382,7 +410,9 @@ public class Records {
case MONTH:
switch(end) {
- case MONTH: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2; break; // P..M
+ case MONTH:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2;
+ break; // P..M
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -390,10 +420,18 @@ public class Records {
case DAY:
switch(end) {
- case DAY: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2; break; // P..D
- case HOUR: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6; break; // P..DT12H
- case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9; break; // P..DT12H60M
- case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 12 + extraSecondIntervalSize; break; // P..DT12H60M60....S
+ case DAY:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2;
+ break; // P..D
+ case HOUR:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6;
+ break; // P..DT12H
+ case MINUTE:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9;
+ break; // P..DT12H60M
+ case SECOND:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 12 + extraSecondIntervalSize;
+ break; // P..DT12H60M60....S
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -401,9 +439,15 @@ public class Records {
case HOUR:
switch(end) {
- case HOUR: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3; break; // PT..H
- case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6; break; // PT..H60M
- case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9 + extraSecondIntervalSize; break; // PT..H12M60....S
+ case HOUR:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3;
+ break; // PT..H
+ case MINUTE:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6;
+ break; // PT..H60M
+ case SECOND:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9 + extraSecondIntervalSize;
+ break; // PT..H12M60....S
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -411,8 +455,12 @@ public class Records {
case MINUTE:
switch(end) {
- case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3; break; // PT...M
- case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6 + extraSecondIntervalSize; break; // PT..M60....S
+ case MINUTE:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3;
+ break; // PT...M
+ case SECOND:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6 + extraSecondIntervalSize;
+ break; // PT..M60....S
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
@@ -421,7 +469,9 @@ public class Records {
case SECOND:
switch(end) {
- case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3 + extraSecondIntervalSize; break; // PT....S
+ case SECOND:
+ this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3 + extraSecondIntervalSize;
+ break; // PT....S
default:
throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" );
}
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java
index f45edbb8c..34c2fe704 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java
@@ -214,7 +214,7 @@ public class ParquetFilterBuilder extends AbstractExprVisitor<LogicalExpression,
}
if (value.contains(funcHolderExpr)) {
- ValueHolder result ;
+ ValueHolder result;
try {
result = InterpreterEvaluator.evaluateConstantExpr(udfUtilities, funcHolderExpr);
} catch (Exception e) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java
index b5f0ca48a..7a0dbbe45 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java
@@ -167,7 +167,7 @@ public abstract class ParquetPushDownFilter extends StoragePluginOptimizerRule {
return;
}
- RelNode newScan = ScanPrel.create(scan, scan.getTraitSet(), newGroupScan, scan.getRowType());;
+ RelNode newScan = ScanPrel.create(scan, scan.getTraitSet(), newGroupScan, scan.getRowType());
if (project != null) {
newScan = project.copy(project.getTraitSet(), ImmutableList.of(newScan));
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java
index c343d317c..047404017 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java
@@ -78,7 +78,10 @@ public abstract class ColumnReader<V extends ValueVector> {
int currDefLevel;
// variables for a single read pass
- long readStartInBytes = 0, readLength = 0, readLengthInBits = 0, recordsReadInThisIteration = 0;
+ long readStartInBytes = 0;
+ long readLength = 0;
+ long readLengthInBits = 0;
+ long recordsReadInThisIteration = 0;
private ExecutorService threadPool;
volatile boolean isShuttingDown; //Indicate to not submit any new AsyncPageReader Tasks during clear()
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java
index b9480e836..f7239d660 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java
@@ -87,7 +87,7 @@ public class MetadataPathUtils {
ParquetTableMetadata_v3 tableMetadataWithAbsolutePaths, String baseDir) {
List<String> directoriesWithRelativePaths = Lists.newArrayList();
for (String directory : tableMetadataWithAbsolutePaths.getDirectories()) {
- directoriesWithRelativePaths.add(relativize(baseDir, directory)) ;
+ directoriesWithRelativePaths.add(relativize(baseDir, directory));
}
List<ParquetFileMetadata_v3> filesWithRelativePaths = Lists.newArrayList();
for (ParquetFileMetadata_v3 file : tableMetadataWithAbsolutePaths.files) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java
index b2961a580..54992076e 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java
@@ -192,7 +192,7 @@ public class BlockMapBuilder {
final Timer.Context context = metrics.timer(BLOCK_MAP_BUILDER_TIMER).time();
BlockLocation[] blocks;
ImmutableRangeMap<Long,BlockLocation> blockMap;
- blocks = fs.getFileBlockLocations(status, 0 , status.getLen());
+ blocks = fs.getFileBlockLocations(status, 0, status.getLen());
ImmutableRangeMap.Builder<Long, BlockLocation> blockMapBuilder = new ImmutableRangeMap.Builder<Long,BlockLocation>();
for (BlockLocation block : blocks) {
long start = block.getOffset();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java b/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java
index 8729a391b..270841f5b 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java
@@ -56,7 +56,7 @@ public class VectorUtil {
int columnCounter = 0;
for (VectorWrapper<?> vw : va) {
boolean lastColumn = columnCounter == width - 1;
- Object o ;
+ Object o;
try{
o = vw.getValueVector().getAccessor().getObject(row);
} catch (Exception e) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java
index fa746770b..30862efcc 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java
@@ -24,7 +24,7 @@ import org.apache.drill.exec.record.RawFragmentBatch;
public interface DataCollector extends AutoCloseable {
static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(DataCollector.class);
- public boolean batchArrived(int minorFragmentId, RawFragmentBatch batch) throws IOException ;
+ public boolean batchArrived(int minorFragmentId, RawFragmentBatch batch) throws IOException;
public int getOppositeMajorFragmentId();
public RawBatchBuffer[] getBuffers();
public int getTotalIncomingFragments();
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java
index 719f3679d..44b7f53fb 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java
@@ -40,7 +40,7 @@ public class UnlimitedRawBatchBuffer extends BaseRawBatchBuffer<RawFragmentBatch
}
private class UnlimitedBufferQueue implements BufferQueue<RawFragmentBatch> {
- private final LinkedBlockingDeque<RawFragmentBatch> buffer = Queues.newLinkedBlockingDeque();;
+ private final LinkedBlockingDeque<RawFragmentBatch> buffer = Queues.newLinkedBlockingDeque();
@Override
public void addOomBatch(RawFragmentBatch batch) {
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java
index 430589135..f4e460d0e 100644
--- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java
+++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java
@@ -53,7 +53,7 @@ public class ResourceManagerBuilder {
private static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(ResourceManagerBuilder.class);
- private DrillbitContext context ;
+ private DrillbitContext context;
public ResourceManagerBuilder(final DrillbitContext context) {
this.context = context;