Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
11 changes: 3 additions & 8 deletions server/src/main/java/io/druid/client/DirectDruidClient.java
Original file line number Diff line number Diff line change
Expand Up @@ -116,21 +116,16 @@ public class DirectDruidClient<T> implements QueryRunner<T>
private final boolean isSmile;

/**
* Removes the magical fields added by {@link #makeResponseContextForQuery(Query, long)}.
* Removes the magical fields added by {@link #makeResponseContextForQuery()}.
*/
public static void removeMagicResponseContextFields(Map<String, Object> responseContext)
{
responseContext.remove(DirectDruidClient.QUERY_FAIL_TIME);
responseContext.remove(DirectDruidClient.QUERY_TOTAL_BYTES_GATHERED);
}

public static Map<String, Object> makeResponseContextForQuery(Query query, long startTimeMillis)
public static Map<String, Object> makeResponseContextForQuery()
{
final Map<String, Object> responseContext = new ConcurrentHashMap<>();
responseContext.put(
DirectDruidClient.QUERY_FAIL_TIME,
startTimeMillis + QueryContexts.getTimeout(query)
);
responseContext.put(
DirectDruidClient.QUERY_TOTAL_BYTES_GATHERED,
new AtomicLong()
Expand Down Expand Up @@ -199,7 +194,7 @@ public Sequence<T> run(final QueryPlus<T> queryPlus, final Map<String, Object> c

final long requestStartTimeNs = System.nanoTime();

long timeoutAt = ((Long) context.get(QUERY_FAIL_TIME)).longValue();
long timeoutAt = query.getContextValue(QUERY_FAIL_TIME);
long maxScatterGatherBytes = QueryContexts.getMaxScatterGatherBytes(query);
AtomicLong totalBytesGathered = (AtomicLong) context.get(QUERY_TOTAL_BYTES_GATHERED);

Expand Down
7 changes: 2 additions & 5 deletions server/src/main/java/io/druid/server/QueryLifecycle.java
Original file line number Diff line number Diff line change
Expand Up @@ -21,14 +21,14 @@

import com.google.common.base.Strings;
import com.google.common.collect.Iterables;
import io.druid.java.util.emitter.service.ServiceEmitter;
import io.druid.client.DirectDruidClient;
import io.druid.java.util.common.DateTimes;
import io.druid.java.util.common.ISE;
import io.druid.java.util.common.guava.Sequence;
import io.druid.java.util.common.guava.SequenceWrapper;
import io.druid.java.util.common.guava.Sequences;
import io.druid.java.util.common.logger.Logger;
import io.druid.java.util.emitter.service.ServiceEmitter;
import io.druid.query.DruidMetrics;
import io.druid.query.GenericQueryMetricsFactory;
import io.druid.query.Query;
Expand Down Expand Up @@ -249,10 +249,7 @@ public QueryResponse execute()
{
transition(State.AUTHORIZED, State.EXECUTING);

final Map<String, Object> responseContext = DirectDruidClient.makeResponseContextForQuery(
queryPlus.getQuery(),
System.currentTimeMillis()
);
final Map<String, Object> responseContext = DirectDruidClient.makeResponseContextForQuery();

final Sequence res = queryPlus.run(texasRanger, responseContext);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,8 @@

package io.druid.server;

import com.google.common.collect.ImmutableMap;
import io.druid.client.DirectDruidClient;
import io.druid.java.util.common.guava.Sequence;
import io.druid.query.Query;
import io.druid.query.QueryContexts;
Expand All @@ -35,11 +37,13 @@ public class SetAndVerifyContextQueryRunner implements QueryRunner
{
private final ServerConfig serverConfig;
private final QueryRunner baseRunner;
private final long startTimeMillis;

public SetAndVerifyContextQueryRunner(ServerConfig serverConfig, QueryRunner baseRunner)
{
this.serverConfig = serverConfig;
this.baseRunner = baseRunner;
this.startTimeMillis = System.currentTimeMillis();
}

@Override
Expand All @@ -54,12 +58,12 @@ public Sequence run(QueryPlus queryPlus, Map responseContext)
);
}

public static <T, QueryType extends Query<T>> QueryType withTimeoutAndMaxScatterGatherBytes(
public <T, QueryType extends Query<T>> QueryType withTimeoutAndMaxScatterGatherBytes(
final QueryType query,
ServerConfig serverConfig
)
{
return (QueryType) QueryContexts.verifyMaxQueryTimeout(
Query<T> newQuery = QueryContexts.verifyMaxQueryTimeout(
QueryContexts.withMaxScatterGatherBytes(
QueryContexts.withDefaultTimeout(
(Query) query,
Expand All @@ -69,5 +73,6 @@ public static <T, QueryType extends Query<T>> QueryType withTimeoutAndMaxScatter
),
serverConfig.getMaxQueryTimeout()
);
return (QueryType) newQuery.withOverriddenContext(ImmutableMap.of(DirectDruidClient.QUERY_FAIL_TIME, this.startTimeMillis + QueryContexts.getTimeout(newQuery)));
}
}
13 changes: 8 additions & 5 deletions server/src/test/java/io/druid/client/DirectDruidClientTest.java
Original file line number Diff line number Diff line change
Expand Up @@ -20,15 +20,12 @@
package io.druid.client;

import com.fasterxml.jackson.core.JsonProcessingException;
import com.google.common.collect.ImmutableMap;
import com.google.common.collect.Lists;
import com.google.common.collect.Maps;
import com.google.common.util.concurrent.Futures;
import com.google.common.util.concurrent.ListenableFuture;
import com.google.common.util.concurrent.SettableFuture;
import io.druid.java.util.http.client.HttpClient;
import io.druid.java.util.http.client.Request;
import io.druid.java.util.http.client.response.HttpResponseHandler;
import io.druid.java.util.http.client.response.StatusResponseHolder;
import io.druid.client.selector.ConnectionCountServerSelectorStrategy;
import io.druid.client.selector.HighestPriorityTierSelectorStrategy;
import io.druid.client.selector.QueryableDruidServer;
Expand All @@ -39,6 +36,10 @@
import io.druid.java.util.common.StringUtils;
import io.druid.java.util.common.guava.Sequence;
import io.druid.java.util.common.guava.Sequences;
import io.druid.java.util.http.client.HttpClient;
import io.druid.java.util.http.client.Request;
import io.druid.java.util.http.client.response.HttpResponseHandler;
import io.druid.java.util.http.client.response.StatusResponseHolder;
import io.druid.query.Druids;
import io.druid.query.QueryInterruptedException;
import io.druid.query.QueryPlus;
Expand Down Expand Up @@ -165,7 +166,7 @@ public void testRun() throws Exception
serverSelector.addServerAndUpdateSegment(queryableDruidServer2, serverSelector.getSegment());

TimeBoundaryQuery query = Druids.newTimeBoundaryQueryBuilder().dataSource("test").build();

query = query.withOverriddenContext(ImmutableMap.of(DirectDruidClient.QUERY_FAIL_TIME, Long.MAX_VALUE));
Sequence s1 = client1.run(QueryPlus.wrap(query), defaultContext);
Assert.assertTrue(capturedRequest.hasCaptured());
Assert.assertEquals(url, capturedRequest.getValue().getUrl());
Expand Down Expand Up @@ -269,6 +270,7 @@ public void testCancel() throws Exception
serverSelector.addServerAndUpdateSegment(queryableDruidServer1, serverSelector.getSegment());

TimeBoundaryQuery query = Druids.newTimeBoundaryQueryBuilder().dataSource("test").build();
query = query.withOverriddenContext(ImmutableMap.of(DirectDruidClient.QUERY_FAIL_TIME, Long.MAX_VALUE));
cancellationFuture.set(new StatusResponseHolder(HttpResponseStatus.OK, new StringBuilder("cancelled")));
Sequence results = client1.run(QueryPlus.wrap(query), defaultContext);
Assert.assertEquals(HttpMethod.DELETE, capturedRequest.getValue().getMethod());
Expand Down Expand Up @@ -340,6 +342,7 @@ public void testQueryInterruptionExceptionLogMessage() throws JsonProcessingExce
serverSelector.addServerAndUpdateSegment(queryableDruidServer, dataSegment);

TimeBoundaryQuery query = Druids.newTimeBoundaryQueryBuilder().dataSource("test").build();
query = query.withOverriddenContext(ImmutableMap.of(DirectDruidClient.QUERY_FAIL_TIME, Long.MAX_VALUE));
interruptionFuture.set(
new ByteArrayInputStream(
StringUtils.toUtf8("{\"error\":\"testing1\",\"errorMessage\":\"testing2\"}")
Expand Down