Druid v0.20.0 Release Notes

Release Date: 2020-10-17 // over 3 years ago
  • ๐Ÿ“š Apache Druid 0.20.0 contains around 160 new features, bug fixes, performance enhancements, documentation improvements, and additional test coverage from 36 contributors. Refer to the complete list of changes and everything tagged to the milestone for further details.

    # New Features

    # Ingestion

    # Combining InputSource

    ๐Ÿ‘€ A new combining InputSource has been added, allowing the user to combine multiple input sources during ingestion. Please see https://druid.apache.org/docs/0.20.0/ingestion/native-batch.html#combining-input-source for more details.

    #10387

    # Automatically determine numShards for parallel ingestion hash partitioning

    When hash partitioning is used in parallel batch ingestion, it is no longer necessary to specify numShards in the partition spec. Druid can now automatically determine a number of shards by scanning the data in a new ingestion phase that determines the cardinalities of the partitioning key.

    #10419

    # Subtask file count limits for parallel batch ingestion

    ๐Ÿ‘ The size-based splitHintSpec now supports a new maxNumFiles parameter, which limits how many files can be assigned to individual subtasks in parallel batch ingestion.

    The segment-based splitHintSpec used for reingesting data from existing Druid segments also has a new maxNumSegments parameter which functions similarly.

    ๐Ÿ‘€ Please see https://druid.apache.org/docs/0.20.0/ingestion/native-batch.html#split-hint-spec for more details.

    #10243

    # Task slot usage metrics

    ๐Ÿ†• New task slot usage metrics have been added. Please see the entries for the taskSlot metrics at https://druid.apache.org/docs/0.20.0/operations/metrics.html#indexing-service for more details.

    #10379

    # Compaction

    ๐Ÿ‘ # Support for all partitioning schemes for auto-compaction

    ๐Ÿ“š A partitioning spec can now be defined for auto-compaction, allowing users to repartition their data at compaction time. Please see the documentation for the new partitionsSpec property in the compaction tuningConfig for more details:

    ๐Ÿ”ง https://druid.apache.org/docs/0.20.0/configuration/index.html#compaction-tuningconfig

    #10307

    # Auto-compaction status API

    A new coordinator API which shows the status of auto-compaction for a datasource has been added. The new API shows whether auto-compaction is enabled for a datasource, and a summary of how far compaction has progressed.

    โšก๏ธ The web console has also been updated to show this information:

    https://user-images.githubusercontent.com/177816/94326243-9d07e780-ff57-11ea-9f80-256fa08580f0.png

    ๐Ÿ‘€ Please see https://druid.apache.org/docs/latest/operations/api-reference.html#compaction-status for details on the new API, and https://druid.apache.org/docs/latest/operations/metrics.html#coordination for information on new related compaction metrics.

    #10371
    #10438

    # Querying

    # Query segment pruning with hash partitioning

    ๐Ÿ‘ Druid now supports query-time segment pruning (excluding certain segments as read candidates for a query) for hash partitioned segments. This optimization applies when all of the partitionDimensions specified in the hash partition spec during ingestion time are present in the filter set of a query, and the filters in the query filter on discrete values of the partitionDimensions (e.g., selector filters). Segment pruning with hash partitioning is not supported with non-discrete filters such as bound filters.

    0๏ธโƒฃ For existing users with existing segments, you will need to reingest those segments to take advantage of this new feature, as the segment pruning requires a partitionFunction to be stored together with the segments, which does not exist in segments created by older versions of Druid. It is not necessary to specify the partitionFunction explicitly, as the default is the same partition function that was used in prior versions of Druid.

    0๏ธโƒฃ Note that segments created with a default partitionDimensions value (partition by all dimensions + the time column) cannot be pruned in this manner, the segments need to be created with an explicit partitionDimensions.

    #9810
    #10288

    # Vectorization

    ๐Ÿ‘€ To enable vectorization features, please set the druid.query.default.context.vectorizeVirtualColumns property to true or set the vectorize property in the query context. Please see https://druid.apache.org/docs/0.20.0/querying/query-context.html#vectorization-parameters for more information.

    ๐Ÿ‘ # Vectorization support for expression virtual columns

    ๐ŸŽ Expression virtual columns now have vectorization support (depending on the expressions being used), which an results in a 3-5x performance improvement in some cases.

    ๐Ÿ‘€ Please see https://druid.apache.org/docs/0.20.0/misc/math-expr.html#vectorization-support for details on the specific expressions that support vectorization.

    #10388
    #10401
    #10432

    ๐Ÿ‘ # More vectorization support for aggregators

    ๐Ÿ‘ Vectorization support has been added for several aggregation types: numeric min/max aggregators, variance aggregators, ANY aggregators, and aggregators from the druid-histogram extension.

    #10260 - numeric min/max
    #10304 - histogram
    #10338 - ANY
    #10390 - variance

    ๐ŸŽ We've observed about a 1.3x to 1.8x performance improvement in some cases with vectorization enabled for the min, max, and ANY aggregator, and about 1.04x to 1.07x wuth the histogram aggregator.

    # offset parameter for GroupBy and Scan queries

    ๐Ÿ‘€ It is now possible set an offset parameter for GroupBy and Scan queries, which tells Druid to skip a number of rows when returning results. Please see https://druid.apache.org/docs/0.20.0/querying/limitspec.html and https://druid.apache.org/docs/0.20.0/querying/scan-query.html for details.

    #10235
    #10233

    # OFFSET clause for SQL queries

    ๐Ÿ‘€ Druid SQL queries now support an OFFSET clause. Please see https://druid.apache.org/docs/0.20.0/querying/sql.html#offset for details.

    #10279

    # Substring search operators

    Druid has added new substring search operators in its expression language and for SQL queries.

    Please see documentation for CONTAINS_STRING and ICONTAINS_STRING string functions for Druid SQL (https://druid.apache.org/docs/0.20.0/querying/sql.html#string-functions) and documentation for contains_string and icontains_string for the Druid expression language (https://druid.apache.org/docs/0.20.0/misc/math-expr.html#string-functions).

    ๐ŸŽ We've observed about a 2.5x performance improvement in some cases by using these functions instead of STRPOS.

    #10350

    # UNION ALL operator for SQL queries

    ๐Ÿ‘€ Druid SQL queries now support the UNION ALL operator, which fuses the results of multiple queries together. Please see https://druid.apache.org/docs/0.20.0/querying/sql.html#union-all for details on what query shapes are supported by this operator.

    #10324

    0๏ธโƒฃ # Cluster-wide default query context settings

    It is now possible to set cluster-wide default query context properties by adding a configuration of the form druid.query.override.default.context.*, with * replaced by the property name.

    #10208

    # Other features

    ๐Ÿ’ป # Improved retention rules UI

    0๏ธโƒฃ The retention rules UI in the web console has been improved. It now provides suggestions and basic validation in the period dropdown, shows the cluster default rules, and makes editing the default rules more accessible.

    #10226

    # Redis cache extension enhancements

    ๐Ÿ’… The Redis cache extension now supports Redis Cluster, selecting which database is used, connecting to password-protected servers, and period-style configurations for the expiration and timeout properties.

    #10240

    # Disable sending server version in response headers

    It is now possible to disable sending of server version information in Druid's response headers.

    0๏ธโƒฃ This is controlled by a new property druid.server.http.sendServerVersion, which defaults to true.

    #9832

    ๐Ÿ”ง # Specify byte-based configuration properties with units

    ๐Ÿ”ง Druid now supports units for specifying byte-based configuration properties, e.g.:

    druid.server.maxSize=300g
    

    equivalent to

    druid.server.maxSize=300000000000
    

    ๐Ÿ‘€ Please see https://druid.apache.org/docs/0.20.0/configuration/human-readable-byte.html for more details.

    #10203

    ๐Ÿ›  # Bug fixes

    # Fix query correctness issue when historical has no segment timeline

    ๐Ÿ›  Druid 0.20.0 fixes a query correctness issue when a broker issues a query expecting a historical to have certain segments for a datasource, but the historical when queried does not actually have any segments for that datasource (e.g., they were all unloaded before the historical processed the query). Prior to 0.20.0, the query would return successfully but without the results from the segments that were missing in the manner described previously. In 0.20.0, queries will now fail in such situations.

    #10199

    # Fix issue preventing result-level cache from being populated

    ๐Ÿ›  Druid 0.20.0 fixes an issue introduced in 0.19.0 (#10337) which can prevent query caches from being populated when result-level caching is enabled.

    #10341

    # Fix for variance aggregator ordering

    The variance aggregator previously used an incorrect comparator that compared using an aggregator's internal count variable instead of the variance.

    #10340

    # Fix incorrect caching for groupBy queries with limit specs

    ๐Ÿ›  Druid 0.20.0 fixes an issues with groupBy queries and caching, where the limitSpec of the query was not considered in the cache key, leading to potentially incorrect results if queries that are identical except for the limitSpec are issued.

    #10093

    # Fix for stringFirst and stringLast with rollup enabled

    ๐Ÿ‘ป #7243 has been resolved, the stringFirst and stringLast aggregators no longer cause an exception when used during ingestion with rollup enabled.

    #10332

    โฌ†๏ธ # Upgrading to Druid 0.20.0

    ๐Ÿš€ Please be aware of the following considerations when upgrading from 0.19.0 to 0.20.0. If you're updating from an earlier version than 0.19.0, please see the release notes of the relevant intermediate versions.

    0๏ธโƒฃ # Default maxSize

    0๏ธโƒฃ druid.server.maxSize will now default to the sum of maxSize values defined within the druid.segmentCache.locations. The user can still provide a custom value for druid.server.maxSize which will take precedence over the default value.

    #10255

    # Compaction and kill task ID changes

    ๐Ÿ›  Compaction and kill tasks issued by the coordinator will now have their task IDs prefixed by coordinator-issued, while user-issued kill tasks will be prefixed by api-issued.

    #10278

    # New size limits for parallel ingestion split hint specs

    0๏ธโƒฃ The size-based and segment-based splitHintSpec for parallel batch ingestion now apply a default file/segment limit of 1000 per subtask, controlled by the maxNumFiles and maxNumSegments respectively.

    #10243

    # New PostAggregator and AggregatorFactory methods

    โšก๏ธ Users who have developed an extension with custom PostAggregator or AggregatorFactory implementions will need to update their extensions, as these two interfaces have new methods defined in 0.20.0.

    PostAggregator now has a new method:

      ValueType getType();
    

    ๐Ÿ‘ To support type information on PostAggregator, AggregatorFactory also has 2 new methods:

      public abstract ValueType getType();
    
      public abstract ValueType getFinalizedType();
    

    ๐Ÿ‘€ Please see #9638 for more details on the interface changes.

    # New Expr-related methods

    โšก๏ธ Users who have developed an extension with custom Expr implementions will need to update their extensions, as Expr and related interfaces hae changed in 0.20.0. Please see the PR below for details:

    #10401

    # More accurate query/cpu/time metric

    In 0.20.0, the accuracy of the query/cpu/time metric has been improved. Previously, it did not account for certain portions of work during query processing, described in more detail in the following PR:

    #10377

    ๐ŸŒฒ # New audit log service metric columns

    โž• If you are using audit logging, please be aware that new columns have been added to the audit log service metric (comment, remote_address, and created_date). An optional payload column has also been added, which can be enabled by setting druid.audit.manager.includePayloadAsDimensionInMetric to true.

    #10373

    ๐Ÿ”Š # sqlQueryContext in request logs

    ๐ŸŒฒ If you are using query request logging, the request log events will now include the sqlQueryContext for SQL queries.

    #10368

    ๐Ÿ“‡ # Additional per-segment state in metadata store

    ๐Ÿ“‡ Hash-partitioned segments created by Druid 0.20.0 will now have additional partitionFunction data in the metadata store.

    โž• Additionally, compaction tasks will now store additional per-segment information in the metadata store, used for tracking compaction history.

    #10288
    #10413

    # Known issues

    # druid.segmentCache.locationSelectorStrategy injection failure

    ๐Ÿ‘€ Specifying a value for druid.segmentCache.locationSelectorStrategy prevents services from starting due to an injection error. Please see #10348 for more details.

    ๐ŸŒ # Resource leak in web console data sampler

    ๐Ÿ‘€ When a timeout occurs while sampling data in the web console, internal resources created to read from the input source are not properly closed. Please see #10467 for more information.

    # Credits

    ๐Ÿš€ Thanks to everyone who contributed to this release!

    @a2l007
    @abhishekagarwal87
    @abhishekrb19
    @ArvinZheng
    @belugabehr
    @capistrant
    @ccaominh
    ๐Ÿ‘• @clintropolis
    @code-crusher
    @dylwylie
    @fermelone
    @FrankChen021
    @gianm
    @himanshug
    @jihoonson
    @jon-wei
    @josephglanville
    @joykent99
    @kroeders
    @lightghli
    @lkm
    @mans2singh
    @maytasm
    @medb
    @mghosh4
    @nishantmonu51
    @pan3793
    @richardstartin
    @sthetland
    @suneet-s
    @tarunparackal
    @tdt17
    @tourvi
    @vogievetsky
    @wjhypo
    @xiangqiao123
    @xvrl


Previous changes from v0.19.0

  • ๐Ÿ“š Apache Druid 0.19.0 contains around 200 new features, bug fixes, performance enhancements, documentation improvements, and additional test coverage from 51 contributors. Refer to the complete list of changes and everything tagged to the milestone for further details.

    # New Features

    0๏ธโƒฃ # GroupBy and Timeseries vectorized query engines enabled by default

    โœ… Vectorized query engines for GroupBy and Timeseries queries were introduced in Druid 0.16, as an opt in feature. Since then we have extensively tested these engines and feel that the time has come for these improvements to find a wider audience. Note that not all of the query engine is vectorized at this time, but this change makes it so that any query which is eligible to be vectorized will do so. This feature may still be disabled if you encounter any problems by setting druid.query.vectorize to false.

    #10065

    ๐Ÿ‘ # Druid native batch support for Apache Avro Object Container Files

    ๐Ÿ†• New in Druid 0.19.0, native batch indexing now supports Apache Avro Object Container Format encoded files, allowing batch ingestion of Avro data without needing an external Hadoop cluster. Check out the docs for more details

    #9671

    โšก๏ธ # Updated Druid native batch support for SQL databases

    ๐Ÿ‘€ An 'SqlInputSource' has been added in Druid 0.19.0 to work with the new native batch ingestion specifications first introduced in Druid 0.17, deprecating the SqlFirehose. Like the 'SqlFirehose' it currently supports MySQL and PostgreSQL, using the driver from those extensions. This is a relatively low level ingestion task, and the operator must take care to manually ensure that the correct data is ingested, either by specially crafting queries to ensure no duplicate data is ingested for appends, or ensuring that the entire set of data is queried to be replaced when overwriting. See the docs for more operational details.

    #9449

    # Apache Ranger based authorization

    ๐Ÿ“š A new extension in Druid 0.19.0 adds an Authorizer which implements access control for Druid, backed by Apache Ranger. Please see [the extension documentation]((https://druid.apache.org/docs/0.19.0/development/extensions-core/druid-ranger-security.html) and Authentication and Authorization for more information on the basic facilities this extension provides.

    #9579

    ๐Ÿ‘ # Alibaba Object Storage Service support

    ๐Ÿ‘€ A new 'contrib' extension has been added for Alibaba Cloud Object Storage Service (OSS) to provide both deep storage and usage as a batch ingestion input source. Since this is a 'contrib' extension, it will not be packaged by default in the binary distribution, please see community extensions for more details on how to use in your cluster.

    #9898

    ๐Ÿ‘ท # Ingestion worker autoscaling for Google Compute Engine

    Another 'contrib' extension new in 0.19.0 has been added to support ingestion worker autoscaling, which allows a Druid Overlord to provision or terminate worker instances (MiddleManagers or Indexers) whenever there are pending tasks or idle workers, for Google Compute Engine. Unlike the Amazon Web Services ingestion autoscaling extension, which provisions and terminates instances directly without using an Auto Scaling Group, the GCE autoscaler uses Managed Instance Groups to more closely align with how operators are likely to provision their clusters in GCE. Like other 'contrib' extensions, it will not be packaged by default in the binary distribution, please see community extensions for more details on how to use in your cluster.

    #8987

    # REGEXP_LIKE

    A new REGEXP_LIKE function has been added to Druid SQL and native expressions, which behaves similar to LIKE, except using regular expressions for the pattern.

    #9893

    ๐ŸŒ # Web console lookup management improvements

    ๐ŸŒ Druid 0.19 also web console also includes some useful improvements to the lookup table management interface. Creating and editing lookups is now done with a form to accept user input, rather than a raw text editor to enter the JSON spec.

    Screen Shot 2020-04-02 at 1 14 38 AM

    โž• Additionally, clicking the magnifying glass icon next to a lookup will now allow displaying the first 5000 values of that lookup.

    Screen Shot 2020-03-20 at 3 09 24 PM

    #9549
    #9587

    # New Coordinator per datasource 'loadstatus' API

    ๐Ÿ“‡ A coordinator API can make it easier to determine if the latest published segments are available for querying. This is similar to the existing coordinator 'loadstatus' API, but is datasource specific, may specify an interval, and can optionally live refresh the metadata store snapshot to get the latest up to date information. Note that operators should still exercise caution when using this API to query large numbers of segments, especially if forcing a metadata refresh, as it can potentially be a 'heavy' call on large clusters.

    #9965

    ๐Ÿ‘ # Native batch append support for range and hash partitioning

    โฌ‡๏ธ Part bug fix, part new feature, Druid native batch (once again) supports appending new data to existing time chunks when those time chunks were partitioned with 'hash' or 'range' partitioning algorithms. Note that currently the appended segments only support 'dynamic' partitioning, and when rolling back to older versions that these appended segments will not be recognized by Druid after the downgrade. In order to roll back to a previous version, these appended segments should be compacted with the rest of the time chunk in order to have a homogenous partitioning scheme.

    #10033

    ๐Ÿ›  # Bug fixes

    ๐Ÿ‘€ Druid 0.19.0 contains 65 bug fixes, you can see the complete list here.

    # Fix for batch ingested 'dynamic' partitioned segments not becoming queryable atomically

    ๐Ÿ›  Druid 0.19.0 fixes an important query correctness issue, where 'dynamic' partitioned segments produced by a batch ingestion task were not tracking the overall number of partitions. This had the implication that when these segments came online, they did not do so as a complete set, but rather as individual segments, meaning that there would be periods of swapping where results could be queried from an incomplete partition set within a time chunk.

    #10025

    # Fix to allow 'hash' and 'range' partitioned segments with empty buckets to now be queryable

    ๐Ÿ›  Prior to 0.19.0, Druid had a bug when using hash or ranged partitioning where if data skew was such that any of the buckets were 'empty' after ingesting, the partitions would never be recognized as 'complete' and so never become queryable. Druid 0.19.0 fixes this issue by adjusting the schema of the partitioning spec. These changes to the json format should be backwards compatible, however rolling back to a previous version will again make these segments no longer queryable.

    #10012

    # Incorrect balancer behavior

    ๐Ÿ”ง A bug in Druid versions prior to 0.19.0 allowed for (incorrect) coordinator operation in the event druid.server.maxSize was not set. This bug would allow segments to load, and effectively randomly balance them in the cluster (regardless of what balancer strategy was actually configured) if all historicals did not have this value set. This bug has been fixed, but as a result druid.server.maxSize must be set to the sum of the segment cache location sizes for historicals, or else they will not load segments.

    #10070

    โฌ†๏ธ # Upgrading to Druid 0.19.0

    ๐Ÿš€ Please be aware of the following issues when upgrading from 0.18.1 to 0.19.0. If you're updating from an earlier version than 0.18.1, please see the release notes of the relevant intermediate versions.

    # 'druid.server.maxSize' must now be set for Historical servers

    โฌ†๏ธ A Coordinator bug fix as a side-effect now requires druid.server.maxSize to be set for segments to be loaded. While this value should have been set correctly for previous versions, please be sure this value is configured correctly before upgrading your clusters or else segments will not be loaded.

    #10070

    ๐Ÿšš # System tables 'sys.segments' column 'payload' has been removed and replaced with 'dimensions', 'metrics', and 'shardSpec'

    ๐Ÿ›ฐ The removal of the 'payload' column from the sys.segments table should make queries on this table much more efficient, and the most useful fields from this, the list of 'dimensions', 'metrics', and the 'shardSpec', have been split out, and so are still available to devote to processing queries.

    #9883

    0๏ธโƒฃ # Changed default number of segment loading threads

    ๐ŸŽ The druid.segmentCache.numLoadingThreads configuration has had the default value changed from 'number of cores' to 'number of cores' divided by 6. This should make historicals a bit more well behaved out of the box when loading a large number of segments, limiting the impact on query performance.

    #9856

    # Broadcast load rules no longer have 'colocated datasources'

    ๐Ÿ”ง A number of incomplete changes to facilitate more efficient join queries, based on the idea of utilizing broadcast load rules to propagate smaller datasources among the cluster so that join operations can be pushed down to individual segment processing, have been added to 0.19.0. While not a finished feature yet, as part of the changes to make this happen, 'broadcast' load rules no longer have the concept of 'colocated datasources', which would attempt to only broadcast segments to servers that had segments of the configured datasource. This didn't work so well in practice, as it was non-atomic, meaning that the broadcast segments would lag behind loads and drops of the colocated datasource, so we decided to remove it.

    #9971

    ๐Ÿ”ง # Brokers and realtime tasks may now be configured to load segments from 'broadcast' datasources

    ๐Ÿ”ง Another effect of the afforementioned preliminary work to introduce efficient 'broadcast joins', Brokers and realtime indexing tasks will now load segments loaded by 'broadcast' rules, if a segment cache is configured. Since the feature is not complete there is little reason to do this in 0.19.0, and it will not happen unless explicitly configured.

    #9971

    # lpad and rpad function behavior change

    0๏ธโƒฃ The lpad and rpad functions have gone through a slight behavior change in Druids default non-SQL compatible mode, in order to make them behave consistently with PostgreSQL. In the new behavior, if the pad expression is an empty string, then the result will be the (possibly trimmed) original characters, rather than the empty string being treated as a null and coercing the results to null.

    #10006

    # Extensions providing custom Druid expressions are now expected to implement equals and hashCode methods

    A change to the Expr interface in Druid 0.19.0 requires that any extension which provides custom expressions via ExprMacroTable must also implement equals and hashCode methods to function correctly, especially with JOIN queries, which rely on filter and expression analysis for determining how to optimally process a query.

    #9830

    # Known Issues

    ๐Ÿ‘€ For a full list of open issues, please see https://github.com/apache/druid/labels/Bug.

    # Credits

    ๐Ÿš€ Thanks to everyone who contributed to this release!

    @2bethere
    @a-chumagin
    @a2l007
    @abhishekrb19
    @agricenko
    @ahuret
    @alex-plekhanov
    @AlexanderSaydakov
    @awelsh93
    @bolkedebruin
    @calvinhkf
    @capistrant
    @ccaominh
    @chenyuzhi459
    ๐Ÿ‘• @clintropolis
    @damnMeddlingKid
    @danc
    @dylwylie
    @egor-ryashin
    @FrankChen021
    @frnidito
    @Fullstop000
    @gianm
    @harshpreet93
    @jihoonson
    @jon-wei
    @josephglanville
    @kamaci
    @kanibs
    @leerho
    @liujianhuanzz
    @maytasm
    @mcbrewster
    @mghosh4
    @morrifeldman
    @pjain1
    @samarthjain
    @stefanbirkner
    @sthetland
    @suneet-s
    @surekhasaharan
    @tarpdalton
    @viongpanzi
    @vogievetsky
    @willsalz
    @wjhypo
    @xhl0726
    @xiangqiao123
    @xvrl
    @yuanlihan
    @zachjsh