diff options
author | Timothy Farkas <timothyfarkas@apache.org> | 2018-08-01 12:32:36 -0700 |
---|---|---|
committer | Vitalii Diravka <vitalii.diravka@gmail.com> | 2018-08-10 21:01:15 +0300 |
commit | 3e3cd2bb76f6a98ca50798bdaa2d86777c7873ad (patch) | |
tree | 817f3e19414b4ec98c06ffea31d6a08227156fa8 /exec/java-exec/src/main/java/org/apache | |
parent | 9a90164ceedc65f33e3a01d3e55282d19de8009b (diff) |
DRILL-6656: Disallow extra semicolons and multiple statements on the same line.
closes #1415
Diffstat (limited to 'exec/java-exec/src/main/java/org/apache')
67 files changed, 233 insertions, 152 deletions
diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java b/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java index f5556cf9b..9bec3933a 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/ExecConstants.java @@ -49,7 +49,7 @@ public final class ExecConstants { public static final String ZK_REFRESH = "drill.exec.zk.refresh"; public static final String BIT_RETRY_TIMES = "drill.exec.rpc.bit.server.retry.count"; public static final String BIT_RETRY_DELAY = "drill.exec.rpc.bit.server.retry.delay"; - public static final String BIT_TIMEOUT = "drill.exec.bit.timeout" ; + public static final String BIT_TIMEOUT = "drill.exec.bit.timeout"; public static final String SERVICE_NAME = "drill.exec.cluster-id"; public static final String INITIAL_BIT_PORT = "drill.exec.rpc.bit.server.port"; public static final String INITIAL_DATA_PORT = "drill.exec.rpc.bit.server.dataport"; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java b/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java index 84c37ba47..8fcb9e921 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/cache/SerializationDefinition.java @@ -29,8 +29,7 @@ public enum SerializationDefinition { STORAGE_PLUGINS(3003, StoragePlugins.class), FRAGMENT_STATUS(3004, FragmentStatus.class), FRAGMENT_HANDLE(3005, FragmentHandle.class), - PLAN_FRAGMENT(3006, PlanFragment.class) - ; + PLAN_FRAGMENT(3006, PlanFragment.class); public final int id; public final Class<?> clazz; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java index ccd7e3636..bca59dc2e 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DrillClient.java @@ -824,7 +824,7 @@ public class DrillClient implements Closeable, ConnectionThrottle { private class ListHoldingResultsListener implements UserResultsListener { private final Vector<QueryDataBatch> results = new Vector<>(); private final SettableFuture<List<QueryDataBatch>> future = SettableFuture.create(); - private final UserProtos.RunQuery query ; + private final UserProtos.RunQuery query; public ListHoldingResultsListener(UserProtos.RunQuery query) { logger.debug( "Listener created for query \"\"\"{}\"\"\"", query ); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java index ff0d0b247..7458c22dd 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/client/DumpCat.java @@ -180,7 +180,7 @@ public class DumpCat { aggBatchMetaInfo.add(getBatchMetaInfo(vcSerializable)); if (vectorContainer.getRecordCount() == 0) { - emptyBatchNum ++; + emptyBatchNum++; } if (prevSchema != null && !vectorContainer.getSchema().equals(prevSchema)) { @@ -188,7 +188,7 @@ public class DumpCat { } prevSchema = vectorContainer.getSchema(); - batchNum ++; + batchNum++; vectorContainer.zeroVectors(); } @@ -219,7 +219,7 @@ public class DumpCat { VectorAccessibleSerializable vcSerializable = null; - while (input.available() > 0 && batchNum ++ < targetBatchNum) { + while (input.available() > 0 && batchNum++ < targetBatchNum) { vcSerializable = new VectorAccessibleSerializable(DumpCat.allocator); vcSerializable.readFromStream(input); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java b/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java index ebb31f492..f15c01a02 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/compile/sig/CodeGeneratorMethod.java @@ -58,7 +58,7 @@ public class CodeGeneratorMethod implements Iterable<CodeGeneratorArgument> { throw new RuntimeException(String.format("Unexpected number of parameter names %s. Expected %s on method %s.", Arrays.toString(parameterNames), Arrays.toString(types), m.toGenericString())); } arguments = new CodeGeneratorArgument[parameterNames.length]; - for (int i = 0 ; i < parameterNames.length; i++) { + for (int i = 0; i < parameterNames.length; i++) { arguments[i] = new CodeGeneratorArgument(parameterNames[i], types[i]); } exs = m.getExceptionTypes(); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java b/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java index e94d9f859..a8b5f4b4c 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/dotdrill/DotDrillType.java @@ -21,10 +21,10 @@ import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.Path; public enum DotDrillType { - VIEW + VIEW; // ,FORMAT // ,STATS - ; + private final String ending; @@ -70,4 +70,4 @@ public enum DotDrillType { b.append("}.drill"); DOT_DRILL_GLOB = b.toString(); } -}
\ No newline at end of file +} diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java index 4486972ee..373cdf69a 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/EvaluationVisitor.java @@ -778,7 +778,7 @@ public class EvaluationVisitor { setBlock.assign(out.getValue(), JExpr.lit(1)); } else { assert (e == null); - eval.assign(out.getValue(), JExpr.lit(1)) ; + eval.assign(out.getValue(), JExpr.lit(1)); } generator.unNestEvalBlock(); // exit from nested block @@ -841,7 +841,7 @@ public class EvaluationVisitor { setBlock.assign(out.getValue(), JExpr.lit(0)); } else { assert (e == null); - eval.assign(out.getValue(), JExpr.lit(0)) ; + eval.assign(out.getValue(), JExpr.lit(0)); } generator.unNestEvalBlock(); // exit from nested block. diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java index fdc469724..6263dec52 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/SizedJBlock.java @@ -42,7 +42,7 @@ public class SizedJBlock { } public void incCounter() { - this.count ++; + this.count++; } public int getCount() { diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java index 1a5df670f..59638922a 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/DrillAggFuncHolder.java @@ -93,13 +93,13 @@ class DrillAggFuncHolder extends DrillFuncHolder { //Loop through all workspace vectors, to get the minimum of size of all workspace vectors. JVar sizeVar = setupBlock.decl(g.getModel().INT, "vectorSize", JExpr.lit(Integer.MAX_VALUE)); JClass mathClass = g.getModel().ref(Math.class); - for (int id = 0; id < getWorkspaceVars().length; id ++) { + for (int id = 0; id < getWorkspaceVars().length; id++) { if (!getWorkspaceVars()[id].isInject()) { setupBlock.assign(sizeVar,mathClass.staticInvoke("min").arg(sizeVar).arg(g.getWorkspaceVectors().get(getWorkspaceVars()[id]).invoke("getValueCapacity"))); } } - for(int i =0 ; i < getWorkspaceVars().length; i++) { + for (int i = 0; i < getWorkspaceVars().length; i++) { if (!getWorkspaceVars()[i].isInject()) { setupBlock.assign(workspaceJVars[i], JExpr._new(g.getHolderType(getWorkspaceVars()[i].getMajorType()))); } @@ -156,7 +156,7 @@ class DrillAggFuncHolder extends DrillFuncHolder { private JVar[] declareWorkspaceVectors(ClassGenerator<?> g) { JVar[] workspaceJVars = new JVar[getWorkspaceVars().length]; - for(int i =0 ; i < getWorkspaceVars().length; i++){ + for (int i = 0; i < getWorkspaceVars().length; i++) { if (getWorkspaceVars()[i].isInject()) { workspaceJVars[i] = g.declareClassField("work", g.getModel()._ref(getWorkspaceVars()[i].getType())); g.getBlock(BlockType.SETUP).assign(workspaceJVars[i], g.getMappingSet().getIncoming().invoke("getContext").invoke("getManagedBuffer")); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java index 412674f13..4789a48d4 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/Hash32Functions.java @@ -346,7 +346,7 @@ public class Hash32Functions { } } - @FunctionTemplate(names = {"hash", "hash32" ,"hash32AsDouble"}, scope = FunctionScope.SIMPLE, nulls = FunctionTemplate.NullHandling.INTERNAL) + @FunctionTemplate(names = {"hash", "hash32", "hash32AsDouble"}, scope = FunctionScope.SIMPLE, nulls = FunctionTemplate.NullHandling.INTERNAL) public static class NullableTimeHash implements DrillSimpleFunc { @Param NullableTimeHolder in; @Output IntHolder out; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java index 43098486c..c23a1a298 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/StringFunctions.java @@ -498,7 +498,7 @@ public class StringFunctions{ for (int id = input.start; id < input.end; id++) { byte currentByte = input.buffer.getByte(id); - out.buffer.setByte(id - input.start, Character.toLowerCase(currentByte)) ; + out.buffer.setByte(id - input.start, Character.toLowerCase(currentByte)); } } } @@ -529,7 +529,7 @@ public class StringFunctions{ for (int id = input.start; id < input.end; id++) { byte currentByte = input.buffer.getByte(id); - out.buffer.setByte(id - input.start, Character.toUpperCase(currentByte)) ; + out.buffer.setByte(id - input.start, Character.toUpperCase(currentByte)); } } } @@ -711,9 +711,9 @@ public class StringFunctions{ final int charCount = org.apache.drill.exec.expr.fn.impl.StringFunctionUtil.getUTF8CharLength(string.buffer, string.start, string.end); final int charLen; if (length.value > 0) { - charLen = Math.min((int)length.value, charCount); //left('abc', 5) -> 'abc' + charLen = Math.min((int) length.value, charCount); //left('abc', 5) -> 'abc' } else if (length.value < 0) { - charLen = Math.max(0, charCount + (int)length.value) ; // left('abc', -5) ==> '' + charLen = Math.max(0, charCount + (int) length.value); // left('abc', -5) ==> '' } else { charLen = 0; } @@ -830,7 +830,7 @@ public class StringFunctions{ if (j == from.end ) { //find a true match ("from" is not empty), copy entire "to" string to out buffer - for (int k = to.start ; k < to.end; k++) { + for (int k = to.start; k < to.end; k++) { out.buffer.setByte(out.end++, to.buffer.getByte(k)); } @@ -921,7 +921,7 @@ public class StringFunctions{ (currentByte & 0xE0) == 0xC0 || // 2-byte char. First byte is 110xxxxx (currentByte & 0xF0) == 0xE0 || // 3-byte char. First byte is 1110xxxx (currentByte & 0xF8) == 0xF0) { //4-byte char. First byte is 11110xxx - count ++; //Advance the counter, since we find one char. + count++; //Advance the counter, since we find one char. } out.buffer.setByte(out.end++, currentByte); } @@ -1070,7 +1070,7 @@ public class StringFunctions{ (currentByte & 0xE0) == 0xC0 || // 2-byte char. First byte is 110xxxxx (currentByte & 0xF0) == 0xE0 || // 3-byte char. First byte is 1110xxxx (currentByte & 0xF8) == 0xF0) { //4-byte char. First byte is 11110xxx - count ++; //Advance the counter, since we find one char. + count++; //Advance the counter, since we find one char. } out.buffer.setByte(out.end++, currentByte); } @@ -1429,7 +1429,7 @@ public class StringFunctions{ @Override public void eval() { - out.buffer = buffer = buffer.reallocIfNeeded( (left.end - left.start) + (right.end - right.start)); + out.buffer = buffer = buffer.reallocIfNeeded((left.end - left.start) + (right.end - right.start)); out.start = out.end = 0; int id = 0; @@ -1493,7 +1493,7 @@ public class StringFunctions{ @Override public void eval() { - out.buffer = buffer = buffer.reallocIfNeeded( (left.end - left.start) + (right.end - right.start));; + out.buffer = buffer = buffer.reallocIfNeeded((left.end - left.start) + (right.end - right.start)); out.start = out.end = 0; int id = 0; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java index e38489236..1bb2512de 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/impl/conv/JsonConvertTo.java @@ -44,7 +44,8 @@ public class JsonConvertTo { private JsonConvertTo(){} - @FunctionTemplate(names = { "convert_toJSON", "convert_toSIMPLEJSON" } , scope = FunctionScope.SIMPLE, nulls = NullHandling.NULL_IF_NULL, + @FunctionTemplate(names = { "convert_toJSON", "convert_toSIMPLEJSON" }, + scope = FunctionScope.SIMPLE, nulls = NullHandling.NULL_IF_NULL, outputWidthCalculatorType = FunctionTemplate.OutputWidthCalculatorType.CUSTOM_FIXED_WIDTH_DEFUALT) public static class ConvertToJson implements DrillSimpleFunc{ diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java index e127391fe..c07e6ed37 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/expr/fn/registry/FunctionRegistryHolder.java @@ -316,7 +316,7 @@ public class FunctionRegistryHolder { final String functionName = function.getName(); Queue<String> jarFunctions = jar.get(functionName); if (jarFunctions == null) { - jarFunctions = Queues.newConcurrentLinkedQueue();; + jarFunctions = Queues.newConcurrentLinkedQueue(); jar.put(functionName, jarFunctions); } final String functionSignature = function.getSignature(); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java index 51f34a017..da988dec4 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashAggregate.java @@ -104,6 +104,6 @@ public class HashAggregate extends AbstractSingle { public boolean isBufferedOperator(QueryContext queryContext) { // In case forced to use a single partition - do not consider this a buffered op (when memory is divided) return queryContext == null || - 1 < (int)queryContext.getOptions().getOption(ExecConstants.HASHAGG_NUM_PARTITIONS_VALIDATOR) ; + 1 < (int) queryContext.getOptions().getOption(ExecConstants.HASHAGG_NUM_PARTITIONS_VALIDATOR); } } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java index 48d977ea1..146bb7fd6 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/config/HashJoinPOP.java @@ -88,6 +88,6 @@ public class HashJoinPOP extends AbstractJoinPop { public boolean isBufferedOperator(QueryContext queryContext) { // In case forced to use a single partition - do not consider this a buffered op (when memory is divided) return queryContext == null || - 1 < (int)queryContext.getOptions().getOption(ExecConstants.HASHJOIN_NUM_PARTITIONS_VALIDATOR) ; + 1 < (int) queryContext.getOptions().getOption(ExecConstants.HASHJOIN_NUM_PARTITIONS_VALIDATOR); } } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java index 4bbfa05a1..1954c79a7 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggTemplate.java @@ -154,7 +154,13 @@ public abstract class HashAggTemplate implements HashAggregator { private int cycleNum = 0; // primary, secondary, tertiary, etc. private int originalPartition = -1; // the partition a secondary reads from - private static class SpilledPartition { public int spilledBatches; public String spillFile; int cycleNum; int origPartn; int prevOrigPartn; } + private static class SpilledPartition { + public int spilledBatches; + public String spillFile; + int cycleNum; + int origPartn; + int prevOrigPartn; + } private ArrayList<SpilledPartition> spilledPartitionsList; private int operatorId; // for the spill file name @@ -193,7 +199,6 @@ public abstract class HashAggTemplate implements HashAggregator { AVG_OUTPUT_BATCH_BYTES, AVG_OUTPUT_ROW_BYTES, OUTPUT_RECORD_COUNT; - ; @Override public int metricId() { @@ -333,7 +338,7 @@ public abstract class HashAggTemplate implements HashAggregator { is2ndPhase = hashAggrConfig.getAggPhase() == AggPrelBase.OperatorPhase.PHASE_2of2; isTwoPhase = hashAggrConfig.getAggPhase() != AggPrelBase.OperatorPhase.PHASE_1of1; - is1stPhase = isTwoPhase && ! is2ndPhase ; + is1stPhase = isTwoPhase && !is2ndPhase; canSpill = isTwoPhase; // single phase can not spill // Typically for testing - force a spill after a partition has more than so many batches @@ -410,8 +415,8 @@ public abstract class HashAggTemplate implements HashAggregator { updateEstMaxBatchSize(incoming); } // create "reserved memory" and adjust the memory limit down - reserveValueBatchMemory = reserveOutgoingMemory = estValuesBatchSize ; - long newMemoryLimit = allocator.getLimit() - reserveValueBatchMemory - reserveOutgoingMemory ; + reserveValueBatchMemory = reserveOutgoingMemory = estValuesBatchSize; + long newMemoryLimit = allocator.getLimit() - reserveValueBatchMemory - reserveOutgoingMemory; long memAvail = newMemoryLimit - allocator.getAllocatedMemory(); if ( memAvail <= 0 ) { throw new OutOfMemoryException("Too little memory available"); } allocator.setLimit(newMemoryLimit); @@ -458,9 +463,9 @@ public abstract class HashAggTemplate implements HashAggregator { bitsInMask = Integer.bitCount(partitionMask); // e.g. 0x1F -> 5 // Create arrays (one entry per partition) - htables = new HashTable[numPartitions] ; - batchHolders = (ArrayList<BatchHolder>[]) new ArrayList<?>[numPartitions] ; - outBatchIndex = new int[numPartitions] ; + htables = new HashTable[numPartitions]; + batchHolders = (ArrayList<BatchHolder>[]) new ArrayList<?>[numPartitions]; + outBatchIndex = new int[numPartitions]; writers = new Writer[numPartitions]; spilledBatchesCount = new int[numPartitions]; spillFiles = new String[numPartitions]; @@ -486,7 +491,11 @@ public abstract class HashAggTemplate implements HashAggregator { this.batchHolders[i] = new ArrayList<BatchHolder>(); // First BatchHolder is created when the first put request is received. } // Initialize the value vectors in the generated code (which point to the incoming or outgoing fields) - try { htables[0].updateBatches(); } catch (SchemaChangeException sc) { throw new UnsupportedOperationException(sc); }; + try { + htables[0].updateBatches(); + } catch (SchemaChangeException sc) { + throw new UnsupportedOperationException(sc); + } } /** * get new incoming: (when reading spilled files like an "incoming") @@ -689,7 +698,7 @@ public abstract class HashAggTemplate implements HashAggregator { // Either flag buildComplete or handleEmit (or earlyOutput) would cause returning of // the outgoing batch downstream (see innerNext() in HashAggBatch). - buildComplete = true ; // now should go and return outgoing + buildComplete = true; // now should go and return outgoing if ( handleEmit ) { buildComplete = false; // This was not a real NONE - more incoming is expected @@ -939,7 +948,7 @@ public abstract class HashAggTemplate implements HashAggregator { } // Give the current (if already spilled) some priority if ( ! tryAvoidCurr && isSpilled(currPart) && ( currPartSize + 1 >= maxSizeSpilled )) { - maxSizeSpilled = currPartSize ; + maxSizeSpilled = currPartSize; indexMaxSpilled = currPart; } // now find the largest non-spilled partition @@ -948,7 +957,7 @@ public abstract class HashAggTemplate implements HashAggregator { // Use the largest spilled (if found) as a base line, with a factor of 4 if ( indexMaxSpilled > -1 && maxSizeSpilled > 1 ) { indexMax = indexMaxSpilled; - maxSize = 4 * maxSizeSpilled ; + maxSize = 4 * maxSizeSpilled; } for ( int insp = 0; insp < numPartitions; insp++) { if ( ! isSpilled(insp) && maxSize < batchHolders[insp].size() ) { @@ -1159,7 +1168,11 @@ public abstract class HashAggTemplate implements HashAggregator { originalPartition = sp.origPartn; // used for the filename logger.trace("Reading back spilled original partition {} as an incoming",originalPartition); // Initialize .... new incoming, new set of partitions - try { initializeSetup(newIncoming); } catch (Exception e) { throw new RuntimeException(e); } + try { + initializeSetup(newIncoming); + } catch (Exception e) { + throw new RuntimeException(e); + } // update the cycle num if needed // The current cycle num should always be one larger than in the spilled partition if ( cycleNum == sp.cycleNum ) { @@ -1179,7 +1192,7 @@ public abstract class HashAggTemplate implements HashAggregator { return AggIterOutcome.AGG_RESTART; } - partitionToReturn = nextPartitionToReturn ; + partitionToReturn = nextPartitionToReturn; } @@ -1187,7 +1200,7 @@ public abstract class HashAggTemplate implements HashAggregator { int numPendingOutput = currPartition.get(currOutBatchIndex).getNumPendingOutput(); // The following accounting is for logging, metrics, etc. - rowsInPartition += numPendingOutput ; + rowsInPartition += numPendingOutput; if ( ! handlingSpills ) { rowsNotSpilled += numPendingOutput; } else { rowsSpilledReturned += numPendingOutput; } if ( earlyOutput ) { rowsReturnedEarly += numPendingOutput; } @@ -1238,7 +1251,7 @@ public abstract class HashAggTemplate implements HashAggregator { logger.debug("HASH AGG: Finished (early) re-init partition {}, mem allocated: {}", earlyPartition, allocator.getAllocatedMemory()); } outBatchIndex[earlyPartition] = 0; // reset, for next time - earlyOutput = false ; // done with early output + earlyOutput = false; // done with early output } else if ( handleEmit ) { // When returning the last outgoing batch (following an incoming EMIT), then replace OK with EMIT @@ -1290,9 +1303,9 @@ public abstract class HashAggTemplate implements HashAggregator { */ private String getOOMErrorMsg(String prefix) { String errmsg; - if ( !isTwoPhase ) { - errmsg = "Single Phase Hash Aggregate operator can not spill." ; - } else if ( ! canSpill ) { // 2nd phase, with only 1 partition + if (!isTwoPhase) { + errmsg = "Single Phase Hash Aggregate operator can not spill."; + } else if (!canSpill) { // 2nd phase, with only 1 partition errmsg = "Too little memory available to operator to facilitate spilling."; } else { // a bug ? errmsg = prefix + " OOM at " + (is2ndPhase ? "Second Phase" : "First Phase") + ". Partitions: " + numPartitions + @@ -1353,9 +1366,11 @@ public abstract class HashAggTemplate implements HashAggregator { } // right shift hash code for secondary (or tertiary...) spilling - for (int i = 0; i < cycleNum; i++) { hashCode >>>= bitsInMask; } + for (int i = 0; i < cycleNum; i++) { + hashCode >>>= bitsInMask; + } - int currentPartition = hashCode & partitionMask ; + int currentPartition = hashCode & partitionMask; hashCode >>>= bitsInMask; HashTable.PutStatus putStatus = null; long allocatedBeforeHTput = allocator.getAllocatedMemory(); @@ -1398,7 +1413,7 @@ public abstract class HashAggTemplate implements HashAggregator { throw new UnsupportedOperationException("Unexpected schema change", e); } long allocatedBeforeAggCol = allocator.getAllocatedMemory(); - boolean needToCheckIfSpillIsNeeded = allocatedBeforeAggCol > allocatedBeforeHTput ; + boolean needToCheckIfSpillIsNeeded = allocatedBeforeAggCol > allocatedBeforeHTput; // Add an Aggr batch if needed: // diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java index f58be8929..4c54650cf 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/HashAggregator.java @@ -44,7 +44,7 @@ public interface HashAggregator { // For returning results from outputCurrentBatch // OK - batch returned, NONE - end of data, RESTART - call again, EMIT - like OK but EMIT - enum AggIterOutcome { AGG_OK, AGG_NONE, AGG_RESTART , AGG_EMIT } + enum AggIterOutcome { AGG_OK, AGG_NONE, AGG_RESTART, AGG_EMIT } void setup(HashAggregate hashAggrConfig, HashTableConfig htConfig, FragmentContext context, OperatorContext oContext, RecordBatch incoming, HashAggBatch outgoing, LogicalExpression[] valueExprs, List<TypedFieldId> valueFieldIds, TypedFieldId[] keyFieldIds, VectorContainer outContainer, int extraRowBytes) throws SchemaChangeException, IOException, ClassTransformationException; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java index c78e2c01f..7ebce2b4a 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/SpilledRecordbatch.java @@ -135,7 +135,7 @@ public class SpilledRecordbatch implements CloseableRecordBatch { if ( spillStream == null ) { throw new IllegalStateException("Spill stream was null"); - }; + } if ( spillSet.getPosition(spillStream) < 0 ) { HashAggTemplate.logger.warn("Position is {} for stream {}", spillSet.getPosition(spillStream), spillStream.toString()); @@ -155,7 +155,7 @@ public class SpilledRecordbatch implements CloseableRecordBatch { throw UserException.dataReadError(e).addContext("Failed reading from a spill file").build(HashAggTemplate.logger); } - spilledBatches-- ; // one less batch to read + spilledBatches--; // one less batch to read return IterOutcome.OK; } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java index 2a64b9307..23fdcc1d2 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/aggregate/StreamingAggregator.java @@ -49,8 +49,7 @@ public interface StreamingAggregator { RETURN_OUTCOME, CLEANUP_AND_RETURN, UPDATE_AGGREGATOR, - RETURN_AND_RESET - ; + RETURN_AND_RESET; } public abstract void setup(OperatorContext context, RecordBatch incoming, StreamingAggBatch outgoing) throws SchemaChangeException; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java index 8f0b7abcb..8c93f1bb6 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableStats.java @@ -28,10 +28,10 @@ public class HashTableStats { } public void addStats (HashTableStats newStats) { - this.numBuckets += newStats.numBuckets ; - this.numEntries += newStats.numEntries ; - this.numResizing += newStats.numResizing ; - this.resizingTime += newStats.resizingTime ; + this.numBuckets += newStats.numBuckets; + this.numEntries += newStats.numEntries; + this.numResizing += newStats.numResizing; + this.resizingTime += newStats.resizingTime; } } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java index 83b72d7c7..3c418b99c 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/common/HashTableTemplate.java @@ -476,7 +476,7 @@ public abstract class HashTableTemplate implements HashTable { if (tableSize > MAXIMUM_CAPACITY) { tableSize = MAXIMUM_CAPACITY; } - originalTableSize = tableSize ; // retain original size + originalTableSize = tableSize; // retain original size threshold = (int) Math.ceil(tableSize * loadf); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java index cd58bfd5d..365451c02 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/flatten/FlattenTemplate.java @@ -113,7 +113,7 @@ public abstract class FlattenTemplate implements Flattener { } catch (OversizedAllocationException ex) { // unable to flatten due to a soft buffer overflow. split the batch here and resume execution. logger.debug("Reached allocation limit. Splitting the batch at input index: {} - inner index: {} - current completed index: {}", - valueIndexLocal, innerValueIndexLocal, currentInnerValueIndexLocal) ; + valueIndexLocal, innerValueIndexLocal, currentInnerValueIndexLocal); /* * TODO diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java index d4d4f927e..b1ea96f05 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinBatch.java @@ -776,7 +776,7 @@ public class HashJoinBatch extends AbstractBinaryRecordBatch<HashJoinPOP> { for (int ind = 0; ind < currentRecordCount; ind++) { int hashCode = ( cycleNum == 0 ) ? partitions[0].getBuildHashCode(ind) : read_right_HV_vector.getAccessor().get(ind); // get the hash value from the HV column - int currPart = hashCode & partitionMask ; + int currPart = hashCode & partitionMask; hashCode >>>= bitsInMask; // Append the new inner row to the appropriate partition; spill (that partition) if needed partitions[currPart].appendInnerRow(buildBatch.getContainer(), ind, hashCode, buildCalc); // may spill if needed diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java index 46f2fa369..639f757ec 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/HashJoinProbeTemplate.java @@ -162,7 +162,7 @@ public abstract class HashJoinProbeTemplate implements HashJoinProbe { */ private int appendBuild(VectorContainer buildSrcContainer, int buildSrcIndex) { // "- 1" to skip the last "hash values" added column - int lastColIndex = buildSrcContainer.getNumberOfColumns() - 1 ; + int lastColIndex = buildSrcContainer.getNumberOfColumns() - 1; for (int vectorIndex = 0; vectorIndex < lastColIndex; vectorIndex++) { ValueVector destVector = container.getValueVector(vectorIndex).getValueVector(); ValueVector srcVector = buildSrcContainer.getValueVector(vectorIndex).getValueVector(); @@ -292,7 +292,7 @@ public abstract class HashJoinProbeTemplate implements HashJoinProbe { int hashCode = ( cycleNum == 0 ) ? partitions[0].getProbeHashCode(recordsProcessed) : read_left_HV_vector.getAccessor().get(recordsProcessed); - int currBuildPart = hashCode & partitionMask ; + int currBuildPart = hashCode & partitionMask; hashCode >>>= bitsInMask; // Set and keep the current partition (may be used again on subsequent probe calls as diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java index b9745371e..52871e2b8 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/JoinUtils.java @@ -235,7 +235,7 @@ public class JoinUtils { if (currentrel instanceof DrillAggregateRel) { agg = (DrillAggregateRel)currentrel; } else if (currentrel instanceof RelSubset) { - currentrel = ((RelSubset) currentrel).getBest() ; + currentrel = ((RelSubset) currentrel).getBest(); } else if (currentrel instanceof DrillLimitRel) { // TODO: Improve this check when DRILL-5691 is fixed. // The problem is that RelMdMaxRowCount currently cannot be used diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java index ff33e2f3c..0eb15a7d0 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/join/LateralJoinBatch.java @@ -626,8 +626,8 @@ public class LateralJoinBatch extends AbstractBinaryRecordBatch<LateralJoinPOP> if (leftUpstream == EMIT || leftUpstream == OK_NEW_SCHEMA) { break; } else { - logger.debug("Output batch still has some space left, getting new batches from left and right. OutIndex: {}" - , outputIndex); + logger.debug("Output batch still has some space left, getting new batches from left and right. OutIndex: {}", + outputIndex); // Get both left batch and the right batch and make sure indexes are properly set leftUpstream = processLeftBatch(); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java index 034d6c23d..5049b067e 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/partitionsender/PartitionSenderRootExec.java @@ -269,7 +269,7 @@ public class PartitionSenderRootExec extends BaseRootExec { // set up partitioning function final LogicalExpression expr = operator.getExpr(); final ErrorCollector collector = new ErrorCollectorImpl(); - final ClassGenerator<Partitioner> cg ; + final ClassGenerator<Partitioner> cg; cg = CodeGenerator.getRoot(Partitioner.TEMPLATE_DEFINITION, context.getOptions()); cg.getCodeGenerator().plainJavaCapable(true); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java index 02d2f1833..f0816b625 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/unnest/UnnestImpl.java @@ -54,7 +54,7 @@ public class UnnestImpl implements Unnest { private RepeatedValueVector.RepeatedAccessor accessor; private RecordBatch outgoing; - private IntVector rowIdVector ; // Allocated and owned by the UnnestRecordBatch + private IntVector rowIdVector; // Allocated and owned by the UnnestRecordBatch private IntVector.Mutator rowIdVectorMutator; /** diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java index 68b546b95..5d2721a81 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/impl/xsort/managed/SortMemoryManager.java @@ -266,7 +266,7 @@ public class SortMemoryManager { memoryLimit = (configMemoryLimit == 0) ? opMemoryLimit : Math.min(opMemoryLimit, configMemoryLimit); - preferredSpillBatchSize = config.spillBatchSize();; + preferredSpillBatchSize = config.spillBatchSize(); preferredMergeBatchSize = config.mergeBatchSize(); // Initialize the buffer memory limit for the first batch. diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java index acc9556f8..24e270a4f 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ColumnState.java @@ -347,8 +347,7 @@ public abstract class ColumnState { .attribute("addVersion", addVersion) .attribute("state", state) .attributeIdentity("writer", writer) - .attribute("vectorState") - ; + .attribute("vectorState"); vectorState.dump(format); format.endObject(); } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java index c7c6fdc5d..2783c5b30 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/physical/rowSet/impl/ResultSetLoaderImpl.java @@ -795,8 +795,7 @@ public class ResultSetLoaderImpl implements ResultSetLoader, LoaderInternals { .attribute("activeSchemaVersion", activeSchemaVersion) .attribute("harvestSchemaVersion", harvestSchemaVersion) .attribute("pendingRowCount", pendingRowCount) - .attribute("targetRowCount", targetRowCount) - ; + .attribute("targetRowCount", targetRowCount); format.attribute("root"); rootState.dump(format); format.attribute("rootWriter"); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java index b70c764e2..cf0125692 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/AbstractPartitionDescriptor.java @@ -44,7 +44,7 @@ public abstract class AbstractPartitionDescriptor implements PartitionDescriptor * Create sublists of the partition locations, each sublist of size * at most {@link PartitionDescriptor#PARTITION_BATCH_SIZE} */ - protected abstract void createPartitionSublists() ; + protected abstract void createPartitionSublists(); /** * Iterator that traverses over the super list of partition locations and diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java index 216c8d27b..4cd9ed3c5 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/StarColumnHelper.java @@ -65,7 +65,7 @@ public class StarColumnHelper { } public static boolean isPrefixedStarColumn(String fieldName) { - return fieldName.indexOf(PREFIXED_STAR_COLUMN) > 0 ; // the delimiter * starts at none-zero position. + return fieldName.indexOf(PREFIXED_STAR_COLUMN) > 0; // the delimiter * starts at none-zero position. } public static boolean isNonPrefixedStarColumn(String fieldName) { diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java index 10c4738ba..862fb59fd 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillJoinRelBase.java @@ -137,7 +137,7 @@ public abstract class DrillJoinRelBase extends Join implements DrillRelNode { // just to make sure Cartesian Join is more expensive // than Non-Cartesian Join. - final int keySize = 1 ; // assume having 1 join key, when estimate join cost. + final int keySize = 1; // assume having 1 join key, when estimate join cost. final DrillCostBase cost = (DrillCostBase) computeHashJoinCostWithKeySize(planner, keySize, mq).multiplyBy(mulFactor); // Cartesian join row count will be product of two inputs. The other factors come from the above estimated DrillCost. @@ -197,7 +197,7 @@ public abstract class DrillJoinRelBase extends Join implements DrillRelNode { ) * buildRowCount * factor; double cpuCost = joinConditionCost * (probeRowCount) // probe size determine the join condition comparison cost - + cpuCostBuild + cpuCostProbe ; + + cpuCostBuild + cpuCostProbe; DrillCostFactory costFactory = (DrillCostFactory) planner.getCostFactory(); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java index 7d070b6ac..fde3896c1 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillLimitRelBase.java @@ -84,7 +84,7 @@ public abstract class DrillLimitRelBase extends SingleRel implements DrillRelNod @Override public double estimateRowCount(RelMetadataQuery mq) { - int off = offset != null ? RexLiteral.intValue(offset) : 0 ; + int off = offset != null? RexLiteral.intValue(offset): 0; if (fetch == null) { // If estimated rowcount is less than offset return 0 diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java index b7881c2c9..6c1b14330 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/common/DrillProjectRelBase.java @@ -60,7 +60,7 @@ import com.google.common.collect.Lists; * Base class for logical and physical Project implemented in Drill */ public abstract class DrillProjectRelBase extends Project implements DrillRelNode { - private final int nonSimpleFieldCount ; + private final int nonSimpleFieldCount; protected DrillProjectRelBase(Convention convention, RelOptCluster cluster, RelTraitSet traits, RelNode child, List<? extends RexNode> exps, RelDataType rowType) { @@ -128,10 +128,10 @@ public abstract class DrillProjectRelBase extends Project implements DrillRelNod for (RexNode expr : this.getProjects()) { if (expr instanceof RexInputRef) { // Simple Field reference. - cnt ++; + cnt++; } else if (expr instanceof RexCall && expr.accept(complexFieldIdentifer)) { // Complex field with named segments only. - cnt ++; + cnt++; } } return cnt; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java index ba55faeb7..f64cd7708 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/cost/DrillCostBase.java @@ -169,7 +169,7 @@ public class DrillCostBase implements DrillRelOptCost { || (this.io == Double.POSITIVE_INFINITY) || (this.network == Double.POSITIVE_INFINITY) || (this.rowCount == Double.POSITIVE_INFINITY) - || (this.memory == Double.POSITIVE_INFINITY) ; + || (this.memory == Double.POSITIVE_INFINITY); } @Override diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java index c96f4e6eb..7497783e3 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillFilterRel.java @@ -51,7 +51,7 @@ public class DrillFilterRel extends DrillFilterRelBase implements DrillRel { } public static DrillFilterRel create(RelNode child, RexNode condition) { - return new DrillFilterRel(child.getCluster(), child.getTraitSet(), child, condition) ; + return new DrillFilterRel(child.getCluster(), child.getTraitSet(), child, condition); } } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java index 2cd60688c..5f2b02c29 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillOptiq.java @@ -301,10 +301,15 @@ public class DrillOptiq { case "CHAR": castType = Types.required(MinorType.VARCHAR).toBuilder().setPrecision(call.getType().getPrecision()).build(); break; - - case "INTEGER": castType = Types.required(MinorType.INT); break; - case "FLOAT": castType = Types.required(MinorType.FLOAT4); break; - case "DOUBLE": castType = Types.required(MinorType.FLOAT8); break; + case "INTEGER": + castType = Types.required(MinorType.INT); + break; + case "FLOAT": + castType = Types.required(MinorType.FLOAT4); + break; + case "DOUBLE": + castType = Types.required(MinorType.FLOAT8); + break; case "DECIMAL": if (!context.getPlannerSettings().getOptions().getOption(PlannerSettings.ENABLE_DECIMAL_DATA_TYPE_KEY).bool_val) { throw UserException @@ -327,7 +332,9 @@ public class DrillOptiq { case "INTERVAL_YEAR": case "INTERVAL_YEAR_MONTH": - case "INTERVAL_MONTH": castType = Types.required(MinorType.INTERVALYEAR); break; + case "INTERVAL_MONTH": + castType = Types.required(MinorType.INTERVALYEAR); + break; case "INTERVAL_DAY": case "INTERVAL_DAY_HOUR": case "INTERVAL_DAY_MINUTE": @@ -337,11 +344,19 @@ public class DrillOptiq { case "INTERVAL_HOUR_SECOND": case "INTERVAL_MINUTE": case "INTERVAL_MINUTE_SECOND": - case "INTERVAL_SECOND": castType = Types.required(MinorType.INTERVALDAY); break; - case "BOOLEAN": castType = Types.required(MinorType.BIT); break; - case "BINARY": castType = Types.required(MinorType.VARBINARY); break; - case "ANY": return arg; // Type will be same as argument. - default: castType = Types.required(MinorType.valueOf(call.getType().getSqlTypeName().getName())); + case "INTERVAL_SECOND": + castType = Types.required(MinorType.INTERVALDAY); + break; + case "BOOLEAN": + castType = Types.required(MinorType.BIT); + break; + case "BINARY": + castType = Types.required(MinorType.VARBINARY); + break; + case "ANY": + return arg; // Type will be same as argument. + default: + castType = Types.required(MinorType.valueOf(call.getType().getSqlTypeName().getName())); } return FunctionCallFactory.createCast(castType, ExpressionPosition.UNKNOWN, arg); } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java index 293749ba6..69e9452d7 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionAllRule.java @@ -61,7 +61,7 @@ public class DrillUnionAllRule extends RelOptRule { call.transformTo(new DrillUnionRel(union.getCluster(), traits, convertedInputs, union.all, true /* check compatibility */)); } catch (InvalidRelException e) { - tracer.warn(e.toString()) ; + tracer.warn(e.toString()); } } } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java index c94d2b5a9..a5a6e0330 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/logical/DrillUnionRel.java @@ -50,7 +50,7 @@ public class DrillUnionRel extends DrillUnionRelBase implements DrillRel { return new DrillUnionRel(getCluster(), traitSet, inputs, all, false /* don't check compatibility during copy */); } catch (InvalidRelException e) { - throw new AssertionError(e) ; + throw new AssertionError(e); } } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java index ca68a7d1a..ff640a7b6 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/AggPrelBase.java @@ -49,7 +49,7 @@ public abstract class AggPrelBase extends DrillAggregateRelBase implements Prel public enum OperatorPhase {PHASE_1of1, PHASE_1of2, PHASE_2of2} - protected OperatorPhase operPhase = OperatorPhase.PHASE_1of1 ; // default phase + protected OperatorPhase operPhase = OperatorPhase.PHASE_1of1; // default phase protected List<NamedExpression> keys = Lists.newArrayList(); protected List<NamedExpression> aggExprs = Lists.newArrayList(); protected List<AggregateCall> phase2AggCallList = Lists.newArrayList(); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java index 2c043689b..caec426be 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/BroadcastExchangePrel.java @@ -56,7 +56,7 @@ public class BroadcastExchangePrel extends ExchangePrel{ final double inputRows = mq.getRowCount(child); final int rowWidth = child.getRowType().getFieldCount() * DrillCostBase.AVG_FIELD_WIDTH; - final double cpuCost = broadcastFactor * DrillCostBase.SVR_CPU_COST * inputRows ; + final double cpuCost = broadcastFactor * DrillCostBase.SVR_CPU_COST * inputRows; final double networkCost = broadcastFactor * DrillCostBase.BYTE_NETWORK_COST * inputRows * rowWidth * numEndPoints; return new DrillCostBase(inputRows, cpuCost, 0, networkCost); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java index d10021cc7..b250d5a94 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/DrillDistributionTrait.java @@ -96,7 +96,7 @@ public class DrillDistributionTrait implements RelTrait { @Override public int hashCode() { - return fields == null ? type.hashCode() : type.hashCode() | fields.hashCode() << 4 ; + return fields == null? type.hashCode(): type.hashCode() | fields.hashCode() << 4; } @Override @@ -106,7 +106,7 @@ public class DrillDistributionTrait implements RelTrait { } if (obj instanceof DrillDistributionTrait) { DrillDistributionTrait that = (DrillDistributionTrait) obj; - return this.type == that.type && this.fields.equals(that.fields) ; + return this.type == that.type && this.fields.equals(that.fields); } return false; } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java index 19499d67b..ade0fb349 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashAggPrule.java @@ -91,7 +91,7 @@ public class HashAggPrule extends AggPruleBase { createTransformRequest(call, aggregate, input, traits); if (create2PhasePlan(call, aggregate)) { - traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ; + traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL); RelNode convertedInput = convert(input, traits); new TwoPhaseSubset(call, distOnAllKeys).go(aggregate, convertedInput); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java index 2272a9b97..c9be46f9c 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/HashToMergeExchangePrel.java @@ -38,7 +38,7 @@ public class HashToMergeExchangePrel extends ExchangePrel { private final List<DistributionField> distFields; private int numEndPoints = 0; - private final RelCollation collation ; + private final RelCollation collation; public HashToMergeExchangePrel(RelOptCluster cluster, RelTraitSet traitSet, RelNode input, List<DistributionField> fields, diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java index a03c2f5d4..223882cc1 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/SingleMergeExchangePrel.java @@ -40,7 +40,7 @@ import org.apache.drill.exec.server.options.OptionManager; public class SingleMergeExchangePrel extends ExchangePrel { - private final RelCollation collation ; + private final RelCollation collation; public SingleMergeExchangePrel(RelOptCluster cluster, RelTraitSet traitSet, RelNode input, RelCollation collation) { super(cluster, traitSet, input); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java index 85f516a66..99b6cde45 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/StreamAggPrule.java @@ -70,7 +70,7 @@ public class StreamAggPrule extends AggPruleBase { final RelTraitSet singleDistTrait = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL).plus(singleDist); if (create2PhasePlan(call, aggregate)) { - traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ; + traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL); RelNode convertedInput = convert(input, traits); new SubsetTransformer<DrillAggregateRel, InvalidRelException>(call){ @@ -138,7 +138,7 @@ public class StreamAggPrule extends AggPruleBase { // createTransformRequest(call, aggregate, input, traits); if (create2PhasePlan(call, aggregate)) { - traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL) ; + traits = call.getPlanner().emptyTraitSet().plus(Prel.DRILL_PHYSICAL); RelNode convertedInput = convert(input, traits); new SubsetTransformer<DrillAggregateRel, InvalidRelException>(call){ diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java index 43e030072..eddbe4f34 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/explain/NumberingRelWriter.java @@ -84,7 +84,9 @@ class NumberingRelWriter implements RelWriter { s.append(" "); if (id != null && id.opId == 0) { - for(int i =0; i < spacer.get(); i++){ s.append('-');} + for (int i = 0; i < spacer.get(); i++) { + s.append('-'); + } }else{ spacer.spaces(s); } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java index ac491e920..124857823 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/planner/physical/visitor/StarColumnConverter.java @@ -248,7 +248,7 @@ public class StarColumnConverter extends BasePrelVisitor<Prel, Void, RuntimeExce for (String s : names) { if (uniqueNames.contains(s)) { - for (int i = 0; ; i++ ) { + for (int i = 0;; i++) { s = s + i; if (! origNames.contains(s) && ! uniqueNames.contains(s)) { break; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java b/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java index 9bb574ce2..90de9af0e 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/rpc/control/Controller.java @@ -42,7 +42,7 @@ public interface Controller extends AutoCloseable { * @param node * @return */ - public ControlTunnel getTunnel(DrillbitEndpoint node) ; + public ControlTunnel getTunnel(DrillbitEndpoint node); public DrillbitEndpoint start(DrillbitEndpoint partialEndpoint, boolean allowPortHunting) throws DrillbitStartupException; diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java index bf005b12e..22d0c7b50 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/ProfileWrapper.java @@ -65,7 +65,7 @@ public class ProfileWrapper { this.profile = profile; this.id = profile.hasQueryId() ? profile.getQueryId() : QueryIdHelper.getQueryId(profile.getId()); //Generating Operator Name map (DRILL-6140) - String profileTextPlan = profile.hasPlan() ? profile.getPlan() : "" ; + String profileTextPlan = profile.hasPlan()? profile.getPlan(): ""; generateOpMap(profileTextPlan); final List<FragmentWrapper> fragmentProfiles = new ArrayList<>(); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java index 5a2a37b6c..cddb75ac0 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/server/rest/profile/SimpleDurationFormat.java @@ -72,6 +72,6 @@ public class SimpleDurationFormat { return (days > 0 ? days + " day " : "") + ((hours + days) > 0 ? hours + " hr " : "") + ((minutes + hours + days) > 0 ? String.format("%02d min ", minutes) : "") + - seconds + "." + String.format("%03d sec", milliSeconds) ; + seconds + "." + String.format("%03d sec", milliSeconds); } } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java b/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java index 28cbe075b..c3ff0c5db 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/ssl/SSLConfig.java @@ -253,8 +253,7 @@ public abstract class SSLConfig { .append("\n\ttrustStorePassword: ").append(getPrintablePassword(getTrustStorePassword())) .append("\n\thandshakeTimeout: ").append(getHandshakeTimeout()) .append("\n\tdisableHostVerification: ").append(disableHostVerification()) - .append("\n\tdisableCertificateVerification: ").append(disableCertificateVerification()) - ; + .append("\n\tdisableCertificateVerification: ").append(disableCertificateVerification()); } return sb.toString(); } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java index d35743fce..fba80e507 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/json/JsonProcessor.java @@ -51,7 +51,7 @@ public interface JsonProcessor { String msg, Object... args); - public boolean ignoreJSONParseError() ; + public boolean ignoreJSONParseError(); public void setIgnoreJSONParseErrors(boolean ignoreJSONParseErrors); } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java index ef8f861e5..6d52b6d02 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/HeaderBuilder.java @@ -224,7 +224,7 @@ public class HeaderBuilder extends TextOutput { // "col", "col_2", "col_2_2", "col_2_2_2". // No mapping scheme is perfect... - for (int l = 2; ; l++) { + for (int l = 2;; l++) { final String rewritten = header + "_" + l; key = rewritten.toLowerCase(); if (! idents.contains(key)) { diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java index a0043efc3..9ee86ca03 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/easy/text/compliant/TextInput.java @@ -96,7 +96,7 @@ final class TextInput { this.lineSeparator = settings.getNewLineDelimiter(); byte normalizedLineSeparator = settings.getNormalizedNewLine(); Preconditions.checkArgument(input instanceof Seekable, "Text input only supports an InputStream that supports Seekable."); - boolean isCompressed = input instanceof CompressionInputStream ; + boolean isCompressed = input instanceof CompressionInputStream; Preconditions.checkArgument(!isCompressed || startPos == 0, "Cannot use split on compressed stream."); // splits aren't allowed with compressed data. The split length will be the compressed size which means we'll normally end prematurely. diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java index c4f89699a..c684e7a52 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/ischema/Records.java @@ -130,9 +130,15 @@ public class Records { break; // 2. SqlTypeName enumerators whose names (currently) do not match SQL's // values for DATA_TYPE: - case CHAR: this.DATA_TYPE = "CHARACTER"; break; - case VARCHAR: this.DATA_TYPE = "CHARACTER VARYING"; break; - case VARBINARY: this.DATA_TYPE = "BINARY VARYING"; break; + case CHAR: + this.DATA_TYPE = "CHARACTER"; + break; + case VARCHAR: + this.DATA_TYPE = "CHARACTER VARYING"; + break; + case VARBINARY: + this.DATA_TYPE = "BINARY VARYING"; + break; case INTERVAL_YEAR: case INTERVAL_YEAR_MONTH: case INTERVAL_MONTH: @@ -145,7 +151,9 @@ public class Records { case INTERVAL_HOUR_SECOND: case INTERVAL_MINUTE: case INTERVAL_MINUTE_SECOND: - case INTERVAL_SECOND: this.DATA_TYPE = "INTERVAL"; break; + case INTERVAL_SECOND: + this.DATA_TYPE = "INTERVAL"; + break; // 3: SqlTypeName enumerators not yet seen and confirmed or handled. default: logger.warn( "Type not handled explicitly (code needs review): " @@ -212,10 +220,18 @@ public class Records { this.CHARACTER_OCTET_LENGTH = null; // This NUMERIC_PRECISION is in bits since NUMERIC_PRECISION_RADIX is 2. switch ( sqlTypeName ) { - case TINYINT: NUMERIC_PRECISION = 8; break; - case SMALLINT: NUMERIC_PRECISION = 16; break; - case INTEGER: NUMERIC_PRECISION = 32; break; - case BIGINT: NUMERIC_PRECISION = 64; break; + case TINYINT: + NUMERIC_PRECISION = 8; + break; + case SMALLINT: + NUMERIC_PRECISION = 16; + break; + case INTEGER: + NUMERIC_PRECISION = 32; + break; + case BIGINT: + NUMERIC_PRECISION = 64; + break; default: throw new AssertionError( "Unexpected " + sqlTypeName.getClass().getName() + " value " @@ -253,9 +269,15 @@ public class Records { this.CHARACTER_OCTET_LENGTH = null; // This NUMERIC_PRECISION is in bits since NUMERIC_PRECISION_RADIX is 2. switch ( sqlTypeName ) { - case REAL: NUMERIC_PRECISION = 24; break; - case FLOAT: NUMERIC_PRECISION = 24; break; - case DOUBLE: NUMERIC_PRECISION = 53; break; + case REAL: + NUMERIC_PRECISION = 24; + break; + case FLOAT: + NUMERIC_PRECISION = 24; + break; + case DOUBLE: + NUMERIC_PRECISION = 53; + break; default: throw new AssertionError( "Unexpected type " + sqlTypeName + " in approximate-types branch" ); @@ -285,7 +307,9 @@ public class Records { this.INTERVAL_TYPE = null; this.INTERVAL_PRECISION = null; switch(sqlTypeName) { - case DATE: this.COLUMN_SIZE = 10; break;// yyyy-MM-dd + case DATE: + this.COLUMN_SIZE = 10; + break;// yyyy-MM-dd case TIME: this.COLUMN_SIZE = this.DATETIME_PRECISION == 0 ? 8 // HH::mm::ss : 8 + 1 + this.DATETIME_PRECISION; @@ -373,8 +397,12 @@ public class Records { switch(start) { case YEAR: switch(end) { - case YEAR: this.COLUMN_SIZE = INTERVAL_PRECISION + 2; break;// P..Y - case MONTH: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 5; break; // P..Y12M + case YEAR: + this.COLUMN_SIZE = INTERVAL_PRECISION + 2; + break;// P..Y + case MONTH: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 5; + break; // P..Y12M default: throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" ); } @@ -382,7 +410,9 @@ public class Records { case MONTH: switch(end) { - case MONTH: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2; break; // P..M + case MONTH: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2; + break; // P..M default: throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" ); } @@ -390,10 +420,18 @@ public class Records { case DAY: switch(end) { - case DAY: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2; break; // P..D - case HOUR: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6; break; // P..DT12H - case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9; break; // P..DT12H60M - case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 12 + extraSecondIntervalSize; break; // P..DT12H60M60....S + case DAY: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 2; + break; // P..D + case HOUR: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6; + break; // P..DT12H + case MINUTE: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9; + break; // P..DT12H60M + case SECOND: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 12 + extraSecondIntervalSize; + break; // P..DT12H60M60....S default: throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" ); } @@ -401,9 +439,15 @@ public class Records { case HOUR: switch(end) { - case HOUR: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3; break; // PT..H - case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6; break; // PT..H60M - case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9 + extraSecondIntervalSize; break; // PT..H12M60....S + case HOUR: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3; + break; // PT..H + case MINUTE: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6; + break; // PT..H60M + case SECOND: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 9 + extraSecondIntervalSize; + break; // PT..H12M60....S default: throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" ); } @@ -411,8 +455,12 @@ public class Records { case MINUTE: switch(end) { - case MINUTE: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3; break; // PT...M - case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6 + extraSecondIntervalSize; break; // PT..M60....S + case MINUTE: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3; + break; // PT...M + case SECOND: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 6 + extraSecondIntervalSize; + break; // PT..M60....S default: throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" ); } @@ -421,7 +469,9 @@ public class Records { case SECOND: switch(end) { - case SECOND: this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3 + extraSecondIntervalSize; break; // PT....S + case SECOND: + this.COLUMN_SIZE = this.INTERVAL_PRECISION + 3 + extraSecondIntervalSize; + break; // PT....S default: throw new AssertionError("Unexpected interval type " + this.INTERVAL_TYPE + " in interval-types branch" ); } diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java index f45edbb8c..34c2fe704 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetFilterBuilder.java @@ -214,7 +214,7 @@ public class ParquetFilterBuilder extends AbstractExprVisitor<LogicalExpression, } if (value.contains(funcHolderExpr)) { - ValueHolder result ; + ValueHolder result; try { result = InterpreterEvaluator.evaluateConstantExpr(udfUtilities, funcHolderExpr); } catch (Exception e) { diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java index b5f0ca48a..7a0dbbe45 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/ParquetPushDownFilter.java @@ -167,7 +167,7 @@ public abstract class ParquetPushDownFilter extends StoragePluginOptimizerRule { return; } - RelNode newScan = ScanPrel.create(scan, scan.getTraitSet(), newGroupScan, scan.getRowType());; + RelNode newScan = ScanPrel.create(scan, scan.getTraitSet(), newGroupScan, scan.getRowType()); if (project != null) { newScan = project.copy(project.getTraitSet(), ImmutableList.of(newScan)); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java index c343d317c..047404017 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/columnreaders/ColumnReader.java @@ -78,7 +78,10 @@ public abstract class ColumnReader<V extends ValueVector> { int currDefLevel; // variables for a single read pass - long readStartInBytes = 0, readLength = 0, readLengthInBits = 0, recordsReadInThisIteration = 0; + long readStartInBytes = 0; + long readLength = 0; + long readLengthInBits = 0; + long recordsReadInThisIteration = 0; private ExecutorService threadPool; volatile boolean isShuttingDown; //Indicate to not submit any new AsyncPageReader Tasks during clear() diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java index b9480e836..f7239d660 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/parquet/metadata/MetadataPathUtils.java @@ -87,7 +87,7 @@ public class MetadataPathUtils { ParquetTableMetadata_v3 tableMetadataWithAbsolutePaths, String baseDir) { List<String> directoriesWithRelativePaths = Lists.newArrayList(); for (String directory : tableMetadataWithAbsolutePaths.getDirectories()) { - directoriesWithRelativePaths.add(relativize(baseDir, directory)) ; + directoriesWithRelativePaths.add(relativize(baseDir, directory)); } List<ParquetFileMetadata_v3> filesWithRelativePaths = Lists.newArrayList(); for (ParquetFileMetadata_v3 file : tableMetadataWithAbsolutePaths.files) { diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java index b2961a580..54992076e 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/store/schedule/BlockMapBuilder.java @@ -192,7 +192,7 @@ public class BlockMapBuilder { final Timer.Context context = metrics.timer(BLOCK_MAP_BUILDER_TIMER).time(); BlockLocation[] blocks; ImmutableRangeMap<Long,BlockLocation> blockMap; - blocks = fs.getFileBlockLocations(status, 0 , status.getLen()); + blocks = fs.getFileBlockLocations(status, 0, status.getLen()); ImmutableRangeMap.Builder<Long, BlockLocation> blockMapBuilder = new ImmutableRangeMap.Builder<Long,BlockLocation>(); for (BlockLocation block : blocks) { long start = block.getOffset(); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java b/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java index 8729a391b..270841f5b 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/util/VectorUtil.java @@ -56,7 +56,7 @@ public class VectorUtil { int columnCounter = 0; for (VectorWrapper<?> vw : va) { boolean lastColumn = columnCounter == width - 1; - Object o ; + Object o; try{ o = vw.getValueVector().getAccessor().getObject(row); } catch (Exception e) { diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java index fa746770b..30862efcc 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/DataCollector.java @@ -24,7 +24,7 @@ import org.apache.drill.exec.record.RawFragmentBatch; public interface DataCollector extends AutoCloseable { static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(DataCollector.class); - public boolean batchArrived(int minorFragmentId, RawFragmentBatch batch) throws IOException ; + public boolean batchArrived(int minorFragmentId, RawFragmentBatch batch) throws IOException; public int getOppositeMajorFragmentId(); public RawBatchBuffer[] getBuffers(); public int getTotalIncomingFragments(); diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java index 719f3679d..44b7f53fb 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/batch/UnlimitedRawBatchBuffer.java @@ -40,7 +40,7 @@ public class UnlimitedRawBatchBuffer extends BaseRawBatchBuffer<RawFragmentBatch } private class UnlimitedBufferQueue implements BufferQueue<RawFragmentBatch> { - private final LinkedBlockingDeque<RawFragmentBatch> buffer = Queues.newLinkedBlockingDeque();; + private final LinkedBlockingDeque<RawFragmentBatch> buffer = Queues.newLinkedBlockingDeque(); @Override public void addOomBatch(RawFragmentBatch batch) { diff --git a/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java b/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java index 430589135..f4e460d0e 100644 --- a/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java +++ b/exec/java-exec/src/main/java/org/apache/drill/exec/work/foreman/rm/ResourceManagerBuilder.java @@ -53,7 +53,7 @@ public class ResourceManagerBuilder { private static final org.slf4j.Logger logger = org.slf4j.LoggerFactory.getLogger(ResourceManagerBuilder.class); - private DrillbitContext context ; + private DrillbitContext context; public ResourceManagerBuilder(final DrillbitContext context) { this.context = context; |