Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,6 @@
*/
package org.apache.beam.sdk.runners;

import static org.apache.beam.sdk.util.CoderUtils.encodeToByteArray;
import static org.apache.beam.sdk.util.SerializableUtils.serializeToByteArray;
import static org.apache.beam.sdk.util.StringUtils.byteArrayToJsonString;
import static org.apache.beam.sdk.util.StringUtils.jsonStringToByteArray;
Expand All @@ -34,7 +33,6 @@
import org.apache.beam.sdk.Pipeline;
import org.apache.beam.sdk.Pipeline.PipelineVisitor;
import org.apache.beam.sdk.coders.Coder;
import org.apache.beam.sdk.coders.CoderException;
import org.apache.beam.sdk.coders.IterableCoder;
import org.apache.beam.sdk.io.BigQueryIO;
import org.apache.beam.sdk.io.PubsubIO;
Expand All @@ -47,7 +45,6 @@
import org.apache.beam.sdk.runners.dataflow.ReadTranslator;
import org.apache.beam.sdk.transforms.AppliedPTransform;
import org.apache.beam.sdk.transforms.Combine;
import org.apache.beam.sdk.transforms.Create;
import org.apache.beam.sdk.transforms.DoFn;
import org.apache.beam.sdk.transforms.Flatten;
import org.apache.beam.sdk.transforms.GroupByKey;
Expand Down Expand Up @@ -844,45 +841,6 @@ private <K, InputT, OutputT> void translateHelper(
}
});

registerTransformTranslator(
Create.Values.class,
new TransformTranslator<Create.Values>() {
@Override
public void translate(
Create.Values transform,
TranslationContext context) {
createHelper(transform, context);
}

private <T> void createHelper(
Create.Values<T> transform,
TranslationContext context) {
context.addStep(transform, "CreateCollection");

Coder<T> coder = context.getOutput(transform).getCoder();
List<CloudObject> elements = new LinkedList<>();
for (T elem : transform.getElements()) {
byte[] encodedBytes;
try {
encodedBytes = encodeToByteArray(coder, elem);
} catch (CoderException exn) {
// TODO: Put in better element printing:
// truncate if too long.
throw new IllegalArgumentException(
"Unable to encode element '" + elem + "' of transform '" + transform
+ "' using coder '" + coder + "'.",
exn);
}
String encodedJson = byteArrayToJsonString(encodedBytes);
assert Arrays.equals(encodedBytes,
jsonStringToByteArray(encodedJson));
elements.add(CloudObject.forString(encodedJson));
}
context.addInput(PropertyNames.ELEMENT, elements);
context.addValueOnlyOutput(PropertyNames.OUTPUT, context.getOutput(transform));
}
});

registerTransformTranslator(
Flatten.FlattenPCollectionList.class,
new TransformTranslator<Flatten.FlattenPCollectionList>() {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,6 +21,7 @@

import static org.hamcrest.Matchers.containsInAnyOrder;
import static org.hamcrest.Matchers.containsString;
import static org.hamcrest.Matchers.hasItem;
import static org.hamcrest.Matchers.instanceOf;
import static org.hamcrest.Matchers.startsWith;
import static org.hamcrest.collection.IsIterableContainingInOrder.contains;
Expand Down Expand Up @@ -840,9 +841,16 @@ public void testApplyIsScopedToExactClass() throws IOException {
CompositeTransformRecorder recorder = new CompositeTransformRecorder();
p.traverseTopologically(recorder);

assertThat("Expected to have seen CreateTimestamped composite transform.",
// The recorder will also have seen a Create.Values composite as well, but we can't obtain that
// transform.
assertThat(
"Expected to have seen CreateTimestamped composite transform.",
recorder.getCompositeTransforms(),
Matchers.<PTransform<?, ?>>contains(transform));
hasItem(transform));
assertThat(
"Expected to have two composites, CreateTimestamped and Create.Values",
recorder.getCompositeTransforms(),
hasItem(Matchers.<PTransform<?, ?>>isA((Class) Create.Values.class)));
}

@Test
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -751,7 +751,7 @@ public void testToSingletonTranslation() throws Exception {
assertEquals(2, steps.size());

Step createStep = steps.get(0);
assertEquals("CreateCollection", createStep.getKind());
assertEquals("ParallelRead", createStep.getKind());

Step collectionToSingletonStep = steps.get(1);
assertEquals("CollectionToSingleton", collectionToSingletonStep.getKind());
Expand Down Expand Up @@ -783,7 +783,7 @@ public void testToIterableTranslation() throws Exception {
assertEquals(2, steps.size());

Step createStep = steps.get(0);
assertEquals("CreateCollection", createStep.getKind());
assertEquals("ParallelRead", createStep.getKind());

Step collectionToSingletonStep = steps.get(1);
assertEquals("CollectionToSingleton", collectionToSingletonStep.getKind());
Expand Down

This file was deleted.

Original file line number Diff line number Diff line change
Expand Up @@ -30,11 +30,9 @@
import org.apache.beam.sdk.runners.PipelineRunner;
import org.apache.beam.sdk.runners.inprocess.GroupByKeyEvaluatorFactory.InProcessGroupByKeyOnly;
import org.apache.beam.sdk.runners.inprocess.GroupByKeyEvaluatorFactory.InProcessGroupByKeyOverrideFactory;
import org.apache.beam.sdk.runners.inprocess.InProcessCreate.InProcessCreateOverrideFactory;
import org.apache.beam.sdk.runners.inprocess.ViewEvaluatorFactory.InProcessViewOverrideFactory;
import org.apache.beam.sdk.transforms.Aggregator;
import org.apache.beam.sdk.transforms.AppliedPTransform;
import org.apache.beam.sdk.transforms.Create;
import org.apache.beam.sdk.transforms.GroupByKey;
import org.apache.beam.sdk.transforms.PTransform;
import org.apache.beam.sdk.transforms.ParDo;
Expand Down Expand Up @@ -83,7 +81,6 @@ public class InProcessPipelineRunner
private static Map<Class<? extends PTransform>, PTransformOverrideFactory>
defaultTransformOverrides =
ImmutableMap.<Class<? extends PTransform>, PTransformOverrideFactory>builder()
.put(Create.Values.class, new InProcessCreateOverrideFactory())
.put(GroupByKey.class, new InProcessGroupByKeyOverrideFactory())
.put(CreatePCollectionView.class, new InProcessViewOverrideFactory())
.put(AvroIO.Write.Bound.class, new AvroIOShardedWriteFactory())
Expand Down
Loading