Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -27,6 +27,7 @@
import org.apache.druid.collections.bitmap.MutableBitmap;
import org.apache.druid.collections.bitmap.RoaringBitmapFactory;
import org.apache.druid.collections.spatial.ImmutableRTree;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.extendedset.intset.ConciseSetUtils;
import org.apache.druid.query.filter.BitmapIndexSelector;
import org.apache.druid.query.filter.BoundDimFilter;
Expand Down Expand Up @@ -60,6 +61,10 @@
@Measurement(iterations = 10)
public class BoundFilterBenchmark
{
static {
NullHandling.initializeForTests();
}

private static final int START_INT = 1_000_000_000;
private static final int END_INT = ConciseSetUtils.MAX_ALLOWED_INTEGER;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@
package org.apache.druid.benchmark;

import it.unimi.dsi.fastutil.ints.IntArrayList;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.io.Closer;
import org.apache.druid.segment.data.ColumnarInts;
import org.apache.druid.segment.data.CompressedVSizeColumnarIntsSupplier;
Expand Down Expand Up @@ -49,6 +50,10 @@
@State(Scope.Benchmark)
public class CompressedColumnarIntsBenchmark
{
static {
NullHandling.initializeForTests();
}

private IndexedInts uncompressed;
private IndexedInts compressed;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,6 +21,7 @@

import com.google.common.base.Function;
import com.google.common.collect.Iterables;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.io.Closer;
import org.apache.druid.segment.data.ColumnarInts;
import org.apache.druid.segment.data.ColumnarMultiInts;
Expand Down Expand Up @@ -54,6 +55,10 @@
@State(Scope.Benchmark)
public class CompressedVSizeColumnarMultiIntsBenchmark
{
static {
NullHandling.initializeForTests();
}

private ColumnarMultiInts uncompressed;
private ColumnarMultiInts compressed;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@
package org.apache.druid.benchmark;

import com.yahoo.sketches.hll.HllSketch;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.query.aggregation.AggregatorFactory;
import org.apache.druid.query.aggregation.BufferAggregator;
import org.apache.druid.query.aggregation.datasketches.hll.HllSketchMergeAggregatorFactory;
Expand Down Expand Up @@ -54,6 +55,10 @@
@State(Scope.Benchmark)
public class DataSketchesHllBenchmark
{
static {
NullHandling.initializeForTests();
}

private final AggregatorFactory aggregatorFactory = new HllSketchMergeAggregatorFactory(
"hll",
"hll",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -28,6 +28,7 @@
import org.apache.druid.collections.bitmap.MutableBitmap;
import org.apache.druid.collections.bitmap.RoaringBitmapFactory;
import org.apache.druid.collections.spatial.ImmutableRTree;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.query.filter.BitmapIndexSelector;
import org.apache.druid.query.filter.DruidDoublePredicate;
import org.apache.druid.query.filter.DruidFloatPredicate;
Expand Down Expand Up @@ -62,6 +63,10 @@
@Measurement(iterations = 10)
public class DimensionPredicateFilterBenchmark
{
static {
NullHandling.initializeForTests();
}

private static final int START_INT = 1_000_000_000;

private static final DimensionPredicateFilter IS_EVEN = new DimensionPredicateFilter(
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,7 @@
import org.apache.druid.benchmark.datagen.BenchmarkColumnSchema;
import org.apache.druid.benchmark.datagen.BenchmarkSchemaInfo;
import org.apache.druid.benchmark.datagen.SegmentGenerator;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.Intervals;
import org.apache.druid.java.util.common.granularity.Granularities;
import org.apache.druid.java.util.common.guava.Sequence;
Expand Down Expand Up @@ -71,6 +72,10 @@
@OutputTimeUnit(TimeUnit.MILLISECONDS)
public class ExpressionAggregationBenchmark
{
static {
NullHandling.initializeForTests();
}

@Param({"1000000"})
private int rowsPerSegment;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@
import org.apache.druid.benchmark.datagen.BenchmarkColumnSchema;
import org.apache.druid.benchmark.datagen.BenchmarkSchemaInfo;
import org.apache.druid.benchmark.datagen.SegmentGenerator;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.Intervals;
import org.apache.druid.java.util.common.granularity.Granularities;
import org.apache.druid.java.util.common.guava.Sequence;
Expand Down Expand Up @@ -70,6 +71,10 @@
@OutputTimeUnit(TimeUnit.MILLISECONDS)
public class ExpressionSelectorBenchmark
{
static {
NullHandling.initializeForTests();
}

@Param({"1000000"})
private int rowsPerSegment;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -102,6 +102,10 @@
@Measurement(iterations = 25)
public class FilterPartitionBenchmark
{
static {
NullHandling.initializeForTests();
}

@Param({"750000"})
private int rowsPerSegment;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -27,6 +27,7 @@
import org.apache.druid.benchmark.datagen.BenchmarkSchemaInfo;
import org.apache.druid.benchmark.datagen.BenchmarkSchemas;
import org.apache.druid.benchmark.query.QueryBenchmarkUtil;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.data.input.InputRow;
import org.apache.druid.jackson.DefaultObjectMapper;
import org.apache.druid.java.util.common.granularity.Granularities;
Expand Down Expand Up @@ -99,6 +100,10 @@
@Measurement(iterations = 25)
public class FilteredAggregatorBenchmark
{
static {
NullHandling.initializeForTests();
}

@Param({"75000"})
private int rowsPerSegment;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@
package org.apache.druid.benchmark;

import org.apache.commons.math3.distribution.NormalDistribution;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.query.aggregation.histogram.FixedBucketsHistogram;
import org.openjdk.jmh.annotations.Benchmark;
import org.openjdk.jmh.annotations.BenchmarkMode;
Expand All @@ -46,6 +47,10 @@
@OutputTimeUnit(TimeUnit.MICROSECONDS)
public class FixedHistogramAddBenchmark
{
static {
NullHandling.initializeForTests();
}

private static final int LOWER_LIMIT = 0;
private static final int UPPER_LIMIT = 100000;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,7 @@

package org.apache.druid.benchmark;

import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.logger.Logger;
import org.apache.druid.query.aggregation.histogram.FixedBucketsHistogram;
import org.openjdk.jmh.annotations.Benchmark;
Expand Down Expand Up @@ -46,6 +47,10 @@
@OutputTimeUnit(TimeUnit.MICROSECONDS)
public class FixedHistogramBenchmark
{
static {
NullHandling.initializeForTests();
}

private static final Logger log = new Logger(FixedHistogramBenchmark.class);

private static final int LOWER_LIMIT = 0;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@
package org.apache.druid.benchmark;

import com.google.common.base.Supplier;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.FileUtils;
import org.apache.druid.java.util.common.MappedByteBufferHandler;
import org.apache.druid.segment.data.ColumnarFloats;
Expand Down Expand Up @@ -55,6 +56,10 @@
@OutputTimeUnit(TimeUnit.MILLISECONDS)
public class FloatCompressionBenchmark
{
static {
NullHandling.initializeForTests();
}

@Param("floatCompress/")
private static String dirPath;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@
import com.google.common.collect.ImmutableList;
import org.apache.druid.benchmark.datagen.BenchmarkColumnSchema;
import org.apache.druid.benchmark.datagen.BenchmarkColumnValueGenerator;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.logger.Logger;
import org.apache.druid.segment.column.ValueType;
import org.apache.druid.segment.data.ColumnarFloatsSerializer;
Expand All @@ -44,6 +45,10 @@

public class FloatCompressionBenchmarkFileGenerator
{
static {
NullHandling.initializeForTests();
}

private static final Logger log = new Logger(FloatCompressionBenchmarkFileGenerator.class);
public static final int ROW_NUM = 5000000;
public static final List<CompressionStrategy> COMPRESSIONS =
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,6 +21,7 @@

import com.google.common.io.Files;
import com.google.common.primitives.Ints;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.io.smoosh.FileSmoosher;
import org.apache.druid.java.util.common.io.smoosh.SmooshedFileMapper;
import org.apache.druid.segment.data.GenericIndexed;
Expand Down Expand Up @@ -60,6 +61,10 @@
@State(Scope.Benchmark)
public class GenericIndexedBenchmark
{
static {
NullHandling.initializeForTests();
}

public static final int ITERATIONS = 10000;

static final ObjectStrategy<byte[]> BYTE_ARRAY_STRATEGY = new ObjectStrategy<byte[]>()
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -33,6 +33,7 @@
import org.apache.druid.collections.DefaultBlockingPool;
import org.apache.druid.collections.NonBlockingPool;
import org.apache.druid.collections.StupidPool;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.data.input.InputRow;
import org.apache.druid.jackson.DefaultObjectMapper;
import org.apache.druid.java.util.common.granularity.Granularities;
Expand Down Expand Up @@ -108,6 +109,10 @@
@Measurement(iterations = 30)
public class GroupByTypeInterfaceBenchmark
{
static {
NullHandling.initializeForTests();
}

private static final SegmentId Q_INDEX_SEGMENT_ID = SegmentId.dummy("qIndex");

@Param({"4"})
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@
package org.apache.druid.benchmark;

import com.google.common.collect.ImmutableMap;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.data.input.InputRow;
import org.apache.druid.data.input.MapBasedInputRow;
import org.apache.druid.java.util.common.StringUtils;
Expand Down Expand Up @@ -48,6 +49,10 @@
@State(Scope.Benchmark)
public class IncrementalIndexRowTypeBenchmark
{
static {
NullHandling.initializeForTests();
}

private IncrementalIndex incIndex;
private IncrementalIndex incFloatIndex;
private IncrementalIndex incStrIndex;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,7 @@
import org.apache.druid.collections.bitmap.MutableBitmap;
import org.apache.druid.collections.bitmap.RoaringBitmapFactory;
import org.apache.druid.collections.spatial.ImmutableRTree;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.query.filter.BitmapIndexSelector;
import org.apache.druid.query.filter.BoundDimFilter;
import org.apache.druid.query.filter.Filter;
Expand Down Expand Up @@ -62,6 +63,10 @@
@Measurement(iterations = 10)
public class LikeFilterBenchmark
{
static {
NullHandling.initializeForTests();
}

private static final int START_INT = 1_000_000;
private static final int END_INT = 9_999_999;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@
package org.apache.druid.benchmark;

import com.google.common.base.Supplier;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.FileUtils;
import org.apache.druid.java.util.common.MappedByteBufferHandler;
import org.apache.druid.segment.data.ColumnarLongs;
Expand Down Expand Up @@ -55,6 +56,10 @@
@OutputTimeUnit(TimeUnit.MILLISECONDS)
public class LongCompressionBenchmark
{
static {
NullHandling.initializeForTests();
}

@Param("longCompress/")
private static String dirPath;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,6 +22,7 @@
import com.google.common.collect.ImmutableList;
import org.apache.druid.benchmark.datagen.BenchmarkColumnSchema;
import org.apache.druid.benchmark.datagen.BenchmarkColumnValueGenerator;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.logger.Logger;
import org.apache.druid.segment.column.ValueType;
import org.apache.druid.segment.data.ColumnarLongsSerializer;
Expand All @@ -44,6 +45,10 @@

public class LongCompressionBenchmarkFileGenerator
{
static {
NullHandling.initializeForTests();
}

private static final Logger log = new Logger(LongCompressionBenchmarkFileGenerator.class);
public static final int ROW_NUM = 5000000;
public static final List<CompressionStrategy> COMPRESSIONS =
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -26,6 +26,7 @@
import org.apache.druid.benchmark.datagen.BenchmarkSchemas;
import org.apache.druid.benchmark.query.QueryBenchmarkUtil;
import org.apache.druid.collections.StupidPool;
import org.apache.druid.common.config.NullHandling;
import org.apache.druid.data.input.InputRow;
import org.apache.druid.jackson.DefaultObjectMapper;
import org.apache.druid.java.util.common.granularity.Granularities;
Expand Down Expand Up @@ -99,6 +100,10 @@
@Measurement(iterations = 25)
public class TopNTypeInterfaceBenchmark
{
static {
NullHandling.initializeForTests();
}

private static final SegmentId Q_INDEX_SEGMENT_ID = SegmentId.dummy("qIndex");

@Param({"1"})
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,7 @@

package org.apache.druid.benchmark;

import org.apache.druid.common.config.NullHandling;
import org.apache.druid.java.util.common.FileUtils;
import org.apache.druid.java.util.common.MappedByteBufferHandler;
import org.apache.druid.java.util.common.logger.Logger;
Expand Down Expand Up @@ -53,6 +54,10 @@
@OutputTimeUnit(TimeUnit.MILLISECONDS)
public class VSizeSerdeBenchmark
{
static {
NullHandling.initializeForTests();
}

private static final Logger log = new Logger(VSizeSerdeBenchmark.class);
@Param({"500000"})
private int values;
Expand Down
Loading