Age | Commit message (Collapse) | Author |
|
|
|
closes #1627
|
|
Added a compiler option in CMakeLists.txt to support the ISO C++ 2011 standard.
Also, changed the CMake min version to 3.1.3 to match the min version specified in protobuf.
closes #1697
|
|
Modified LateralJoin to use new setCurrentOutgoingMaxRowCount api
Limit CurrentOutgoingMaxRowCount to MAX_NUM_ROWS
Fix HashJoin to fix failing tests
closes #1650
|
|
and allocated size
|
|
Queues.
closes #1677
|
|
does not return values is called
- Fix check for return function value to handle the case when created object is returned without assigning it to the local variable
closes #1687
|
|
generated code to allow scalar replacement for more cases
closes #1686
|
|
references inside if block
|
|
- Added new optimizer rule which checks if query references directory columns only and has DISTINCT or GROUP BY operation. If the condition holds, instead of scanning full file set the following will be performed:
1) if there is cache metadata file, these directories will be read from it,
2) otherwise directories will be gathered from selection object (PartitionLocation).
In the end Scan node will be transformed to DrillValuesRel (containing constant literals) with gathered values so no scan will be performed.
closes #1640
|
|
closes #1666
|
|
Freemarker by default introduces a comma in numeric values greater than 999. This corrects that by removing the ',' in the default limit size.
However, since a Server-side implementation is in progress (DRILL-6960 and DRILL-7048), it is best to disable this for now. The latest commit in this will hide those capabilities in the WebUI until the server-side feature goes in.
X
closes #1689
|
|
Add control file to generate debian package where we have deleted in #1669 PR. We need control file either in static form or dynamic form. The control file gives details about the package and conffiles used when a new version is installed then dpkg knows to ask what do you want to do with differences instead of just replacing file. According to debian packaging rules the control files are must and conffile is optional. Hence to add control file. We also need to add "<exclude>**/control</exclude>" in distribution/pom.xml file. Because without this line "mvn clean install -DskipTests -Drat.skip=false -Dlicense.skip=false" will fail.
Reference link for debian packaging: https://wiki.debian.org/Packaging/Intro?action=show&redirect=IntroDebianPackaging
Signed-off-by: Naresh Bhat <naresh.bhat@linaro.org>
closes #1679
|
|
- Update dependencies versions
- Exclude dependency from jdbc-all
- Removal redundant "bcpkix-jdk15on" exclusion
- Proper exclusion of "jackson-dataformat-hocon" dependency
- Removal redundant "excludeSubprojects" config property for Maven Rat Plugin
closes #1682
|
|
1. Renamed map to struct in schema parser.
2. Updated sqlTypeOf function to return STRUCT instead of MAP, drillTypeOf function will return MAP as before until internal renaming is done.
3. Add is_struct alias to already existing is_map function. Function should be revisited once Drill supports true maps.
4. Updated unit tests.
closes #1688
|
|
|
|
closes #1685
|
|
1. Add format, default, column properties logic.
2. Changed schema JSON after serialization.
3. Added appropriate unit tests.
closes #1684
|
|
The result set loader allows controlling batch sizes. The new scan framework
built on top of that framework handles projection, implicit columns, null
columns and more. This commit converts the "new" ("compliant") text reader
to use the new framework. Options select the use of the V2 ("new") or V3
(row-set based) versions. Unit tests demonstrate V3 functionality.
closes #1683
|
|
closes #1678
- added CircleCI Test Summary
- maven-surefire-plugin version udpate
- license checkstyle goal is added to the TravisCI protobuf phase
|
|
destruction due to communication error
closes #1660
|
|
closes #1652
|
|
closes #1665
|
|
port is used
closes #1656
|
|
closes #1667
|
|
closes #1674
|
|
query textbox on Query tab
|
|
- replaced all String path representation with org.apache.hadoop.fs.Path
- added PathSerDe.Se JSON serializer
- refactoring of DFSPartitionLocation code by leveraging existing listPartitionValues() functionality
closes #1657
|
|
Roll-up of fixes an enhancements that emerged from the effort to host the CSV reader on the new framework.
closes #1676
|
|
1. Updated protobuf to version 3.6.1
2. Added protobuf to the root pom dependency management
3. Added classes BoundedByteString and LiteralByteString for compatibility with HBase
4. Added ProtobufPatcher to provide compatibility with MapR-DB and HBase
closes #1639
|
|
'ALLOW_BACKSLASH_ESCAPING_ANY_CHARACTER' (#1663)
|
|
The Apache Drill failed to generate debian and rpm packages. The issues fixed as below,
-Add apache as prefix
-Remove lib folder while packaging, because lib folder does not exist.
-Add git.properties, KEYS, LICENSE, NOTICE, README.md and sample-data dir into the /opt/drill/.
-Update plugin version to 2.2.0
-Update Maven org.vafer:jdeb plugin version to 1.7
-Change the rpm name tobe generated by editing version, and name.
-Remove two files under distribution/src/deb/control/ i.e. conffiles and control.
-The `<copyright>` tag isn't supported anymore for this plugin. Hence use `<license>` and project variables content instead of `2013 ASF`.
Signed-off-by: Naresh Bhat <naresh.bhat@linaro.org>
|
|
parquet reader is used
closes #1655
|
|
implemented
- Implemented 'repeated_count' function for repeated MAP and repeated LIST;
- Updated RepeatedListReader and RepeatedMapReader implementations to return correct value from size() method
- Moved repeated_count to freemarker template and added support for more repeated types for the function
closes #1641
|
|
closes #1661
|
|
- Made `NullExpression`s in `IfExpression` with nested `IfExpression`s to be rewritten to typed ones recursively if necessary
closes #1668
|
|
|
|
Add support for avg row-width and major type statistics.
Parallelize the ANALYZE implementation and stats UDF implementation to improve stats collection performance.
Update/fix rowcount, selectivity and ndv computations to improve plan costing.
Add options for configuring collection/usage of statistics.
Add new APIs and implementation for stats writer (as a precursor to Drill Metastore APIs).
Fix several stats/costing related issues identified while running TPC-H nad TPC-DS queries.
Add support for CPU sampling and nested scalar columns.
Add more testcases for collection and usage of statistics and fix remaining unit/functional test failures.
Thanks to Venki Korukanti (@vkorukanti) for the description below (modified to account for new changes). He graciously agreed to rebase the patch to latest master, fixed few issues and added few tests.
FUNCS: Statistics functions as UDFs:
Separate
Currently using FieldReader to ensure consistent output type so that Unpivot doesn't get confused. All stats columns should be Nullable, so that stats functions can return NULL when N/A.
* custom versions of "count" that always return BigInt
* HyperLogLog based NDV that returns BigInt that works only on VarChars
* HyperLogLog with binary output that only works on VarChars
OPS: Updated protobufs for new ops
OPS: Implemented StatisticsMerge
OPS: Implemented StatisticsUnpivot
ANALYZE: AnalyzeTable functionality
* JavaCC syntax more-or-less copied from LucidDB.
* (Basic) AnalyzePrule: DrillAnalyzeRel -> UnpivotPrel StatsMergePrel FilterPrel(for sampling) StatsAggPrel ScanPrel
ANALYZE: Add getMetadataTable() to AbstractSchema
USAGE: Change field access in QueryWrapper
USAGE: Add getDrillTable() to DrillScanRelBase and ScanPrel
* since ScanPrel does not inherit from DrillScanRelBase, this requires adding a DrillTable to the constructor
* This is done so that a custom ReflectiveRelMetadataProvider can access the DrillTable associated with Logical/Physical scans.
USAGE: Attach DrillStatsTable to DrillTable.
* DrillStatsTable represents the data scanned from a corresponding ".stats.drill" table
* In order to avoid doing query execution right after the ".stats.drill" table is found, metadata is not actually collected until the MaterializationVisitor is used.
** Currently, the metadata source must be a string (so that a SQL query can be created). Doing this with a table is probably more complicated.
** Query is set up to extract only the most recent statistics results for each column.
closes #729
|
|
|
|
Adds the "plumbing" that connects the scan operator to the result set loader and the scan projection framework. See the various package-info.java files for the technical datails. Also adds a large number of tests.
This PR does not yet introduce an actual scan operator: that will follow in subsequent PRs.
closes #1618
|
|
closes #1644
This PR standardizes error and alert messages to a cleaner interface by leveraging Bootstraps UX elements for publishing the messages in a presentable format.
Exceptions reported back to the browser and rendered in a neat tabular format (using Panels)
All errors can be redirected to errorMessage.ftl which will render it in a neat format.
Alerts are replaced with modals.
Interactions (pages) affected by Alert modals
1. Missing Query submission
2. profile Query Rerun
3. invalid Profile Listing Fetch
4. invalid Option Value for update
5. Missing username/password submission
The errorMessage.ftl has been moved to root dir, and unused `error.ftl` was removed
|
|
closes #1642
- Add output column names to JdbcRecordReader and use them for storing the results since column names in result set may differ when aliases aren't specified
|
|
closes #1530
|
|
metadata auto-refresh
closes #1638
|
|
closes #1626
|
|
DRILL-7006 added a type conversion "shim" within the row set framework. Basically, we insert a "shim" column writer that takes data in one form (String, say), and does reader-specific conversions to a target format (INT, say).
The code works fine, but the shim class ends up needing to override a bunch of methods which it then passes along to the base writer. This PR refactors the code so that the conversion shim is simpler.
closes #1633
|
|
closes #1632
|
|
if all protobufs are up-to-date
closes #1636
|
|
1. Moved Calcite dependency from profile hadoop-default to general dependency managment
2. Updated Calcite version to 1.18.0-drill-r0 and Avatica version to 1.13.0
3. Hook.REL_BUILDER_SIMPLIFY moved to static block, cause now it can't be removed (fixes DRILL-6830)
4. Removed WrappedAccessor, since it was workaround fixed in CALCITE-1408
5. Fixed setting of multiple options in TestBuilder
6. Timstampadd type inference aligned with CALCITE-2699
7. Dependency update caused 417 kB increase of jdb-all jar size, so the maxsize limit was
increased from 39.5 to 40 MB
8. Added test into TestDrillParquetReader to ensure that DRILL-6856 was
fixed by Calcite update
close apache/drill#1631
|
|
and filter condition is swapped
close apache/drill#1628
|