From 2eb1cf47227a72b25fd644a61bf4231deea94509 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Mon, 20 Jan 2020 13:58:25 +0000 Subject: [PATCH 01/23] [Relay] Added MergeComposite pass This pass allows for patterns to be wrapped in a function marked with 'Composite' and a composite function name. This is intended to be used with the external codegen for the cases where an external operator maps to multiple Relay operators. In that case, the mapping can be expressed as a pattern and assigned a name. For more information on this pass and its motivation, see the RFC: https://discuss.tvm.ai/t/rfc-external-codegen-defining-composite-relay-operators/5470 Change-Id: Icb1b803a9f0ac57c529143200228f3bb5793afc0 --- include/tvm/relay/expr.h | 2 + python/tvm/relay/transform.py | 17 +++ src/relay/pass/merge_composite.cc | 192 ++++++++++++++++++++++++++++++ 3 files changed, 211 insertions(+) create mode 100644 src/relay/pass/merge_composite.cc diff --git a/include/tvm/relay/expr.h b/include/tvm/relay/expr.h index 64f2278c3103..1dcf957426c0 100644 --- a/include/tvm/relay/expr.h +++ b/include/tvm/relay/expr.h @@ -561,6 +561,8 @@ constexpr const char* kParams = "__params__"; constexpr const char* kExternalSymbol = "ExternalSymbol"; /*! \brief Mark if the function should be avoided being optimized. */ constexpr const char* kSkipOptimization = "SkipOptimization"; +/*! \brief Treat the function as a composite operator. */ +constexpr const char* kComposite = "Composite"; } // namespace attr } // namespace relay diff --git a/python/tvm/relay/transform.py b/python/tvm/relay/transform.py index 26b20e01c623..d76727732842 100644 --- a/python/tvm/relay/transform.py +++ b/python/tvm/relay/transform.py @@ -513,6 +513,23 @@ def Legalize(legalize_map_attr_name="FTVMLegalize"): return _transform.Legalize(legalize_map_attr_name) +def MergeComposite(compiler): + """Merge multiple operators into a single composite relay function. + + Parameters + ---------- + compiler : str + The compiler used for codegen. + + Returns + ------- + ret : tvm.relay.Pass + The registered pass that merges operators into a single composite + relay function. + """ + return _transform.MergeComposite(compiler) + + def RewriteAnnotatedOps(fallback_device): """Rewrite the annotated program where annotation operators, e.g. `on_deivce`, mark which device an expression should be scheduled to. diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc new file mode 100644 index 000000000000..81d50b91a0b7 --- /dev/null +++ b/src/relay/pass/merge_composite.cc @@ -0,0 +1,192 @@ +/* + * Licensed to the Apache Software Foundation (ASF) under one + * or more contributor license agreements. See the NOTICE file + * distributed with this work for additional information + * regarding copyright ownership. The ASF licenses this file + * to you under the Apache License, Version 2.0 (the + * "License"); you may not use this file except in compliance + * with the License. You may obtain a copy of the License at + * + * http://www.apache.org/licenses/LICENSE-2.0 + * + * Unless required by applicable law or agreed to in writing, + * software distributed under the License is distributed on an + * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY + * KIND, either express or implied. See the License for the + * specific language governing permissions and limitations + * under the License. + */ + +/*! + * \file src/relay/pass/merge_composite.cc + * \brief Merges expressions matching patterns into functions marked + * as 'composite'. + */ + +#include +#include +#include +#include +#include + +namespace tvm { +namespace relay { +namespace merge_composite { + + +class MergeCompositeWrapper : public ExprMutator { + public: + explicit MergeCompositeWrapper(const tvm::Map& pattern_map) + : pattern_map_(pattern_map) {} + + bool MatchPattern(const Call& pattern, const Call& root) { + if (!pattern->op->IsInstance() || !root->op->IsInstance()) + return false; + if (pattern->op.as()->name != root->op.as()->name) + return false; + if (pattern->args.size() != root->args.size()) + return false; + + unsigned int i = 0; + for (const auto& arg : pattern->args) { + if (arg->IsInstance()) { + if (!root->args[i]->IsInstance()) + return false; + if (!MatchPattern(Downcast(arg), Downcast(root->args[i]))) + return false; + } + i++; + } + return true; + } + + Expr ExtractPattern(const Var& pattern, const Expr& root, + Map>* var_map) { + if (var_map->find(pattern->name_hint()) == var_map->end()) { + auto free_var = VarNode::make(pattern->name_hint(), Type()); + var_map->Set(pattern->name_hint(), Array({free_var, root})); + return std::move(free_var); + } else { + return (*var_map)[pattern->name_hint()][0]; + } + } + + Expr ExtractPattern(const Constant& pattern, const Expr& root, + Map>* var_map) { + return root; + } + + Expr ExtractPattern(const Call& pattern, const Call& root, + Map>* var_map) { + Expr expr; + Expr empty_expr; + if (!pattern->op->IsInstance() || !root->op->IsInstance()) + return empty_expr; + if (pattern->op.as()->name != root->op.as()->name) + return empty_expr; + if (pattern->args.size() != root->args.size()) + return empty_expr; + + unsigned int i = 0; + Array new_args; + for (const auto& arg : pattern->args) { + if (arg->IsInstance()) { + new_args.push_back(ExtractPattern(Downcast(arg), + Downcast(root->args[i]), + var_map)); + } + if (arg->IsInstance()) { + new_args.push_back(ExtractPattern(Downcast(arg), + root->args[i], + var_map)); + } + if (arg->IsInstance()) { + new_args.push_back(ExtractPattern(Downcast(arg), + root->args[i], + var_map)); + } + i++; + } + + auto new_call = CallNode::make(root->op, new_args, root->attrs); + return std::move(new_call); + } + + Expr VisitExpr_(const CallNode* cn) { + Call call = GetRef(cn); + if (call->op->IsInstance()) { + Function func = Downcast(call->op); + CHECK(func.defined()); + const auto name_node = FunctionGetAttr(func, attr::kComposite).as(); + if (name_node->value != "") { + tvm::Array new_args; + for (const auto& arg : call->args) { + auto new_e = this->Mutate(arg); + new_args.push_back(new_e); + } + return CallNode::make(call->op, new_args, call->attrs); + } + } + + Expr expr = ExprMutator::VisitExpr_(cn); + call = Downcast(expr); + if (!call->op->IsInstance()) + return std::move(call); + + Op op = Downcast(call->op); + CHECK(op.defined()); + for (const auto& x : pattern_map_) { + Call pattern = Downcast(x.second); + if (Downcast(pattern->op)->name != op->name) + continue; + + if (MatchPattern(pattern, call)) { + Map> args_map; + auto extract = ExtractPattern(pattern, call, &args_map); + auto free_vars = FreeVars(extract); + Function new_func = FunctionNode::make(free_vars, extract, + call->checked_type_, {}, Attrs()); + new_func = FunctionSetAttr(new_func, attr::kComposite, + tir::StringImmNode::make(x.first)); + new_func = FunctionSetAttr(new_func, attr::kPrimitive, + tvm::Integer(1)); + Array args; + for (const auto& free_var : free_vars) { + args.push_back(args_map[free_var->name_hint()][1]); + } + auto new_call = CallNode::make(new_func, args); + return std::move(new_call); + } + } + + return std::move(call); + } + + private: + tvm::Map pattern_map_; +}; + +Expr MergeComposite(const Expr& expr, const tvm::Map& pattern) { + return MergeCompositeWrapper(pattern).Mutate(expr); +} + +} // namespace merge_composite + +namespace transform { + +Pass MergeComposite(const tvm::Map& pattern) { + runtime::TypedPackedFunc pass_func = + [=](Function f, IRModule m, PassContext pc) { + return Downcast(relay::merge_composite::MergeComposite(f, pattern)); + }; + auto func_pass = CreateFunctionPass(pass_func, 0, "MergeComposite", {}); + return func_pass; +} + +TVM_REGISTER_GLOBAL("relay._transform.MergeComposite") +.set_body_typed(MergeComposite); + +} // namespace transform + +} // namespace relay +} // namespace tvm From 81906dc9c966b30510a65b8c5a7b5434d48f1077 Mon Sep 17 00:00:00 2001 From: Luke Hutton Date: Wed, 22 Jan 2020 14:13:24 +0000 Subject: [PATCH 02/23] [Relay] Merge composite tests Added tests for the merge_composite pass. Change-Id: I1728b4a05b0c1c36140a40f1afe028fde62185dd --- .../python/relay/test_pass_merge_composite.py | 158 ++++++++++++++++++ 1 file changed, 158 insertions(+) create mode 100644 tests/python/relay/test_pass_merge_composite.py diff --git a/tests/python/relay/test_pass_merge_composite.py b/tests/python/relay/test_pass_merge_composite.py new file mode 100644 index 000000000000..e44b18779f42 --- /dev/null +++ b/tests/python/relay/test_pass_merge_composite.py @@ -0,0 +1,158 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +"""Unit tests for merge composite.""" +from tvm import relay +from tvm.relay.testing import run_opt_pass + + +def make_add_sub_mul_pattern(): + """Create a pattern to match the following graph. + + add sub + \ / + \ / + mul + """ + x = relay.var('x') + y = relay.var('y') + add_node = relay.add(x, y) + sub_node = relay.subtract(x, y) + mul_node = relay.multiply(add_node, sub_node) + return mul_node + + +def make_add_relu_pattern(): + """Create a pattern to match the following graph. + + add + | + ReLu + """ + x = relay.var('x') + y = relay.var('y') + add_node = relay.add(x, y) + r = relay.nn.relu(add_node) + return r + + +def test_simple_merge(): + """Test composite function is correctly produced from simple graph. + + We could expect the pattern `make_add_relu_pattern` to be merged + into a single op `add_relu`. + + a b + \ / a b + add ====> \ / + | add_relu + ReLu + + """ + pattern_table = { + "add_sub_mul": make_add_relu_pattern() + } + + def before(): + a = relay.var('a', shape=(10, 10)) + b = relay.var('b', shape=(10, 10)) + add_node = relay.add(a, b) + r = relay.nn.relu(add_node) + return relay.Function([a, b], r) + + def expected(): + a = relay.var('a', shape=(10, 10)) + b = relay.var('b', shape=(10, 10)) + + # add_relu function + in_1 = relay.var('in_1', shape=(10, 10)) + in_2 = relay.var('in_2', shape=(10, 10)) + add_node = relay.add(in_1, in_2) + relu_node = relay.nn.relu(add_node) + add_relu = relay.Function([in_1, in_2], relu_node) + + # merged function + r = relay.Call(add_relu, [a, b]) + return relay.Function([a, b], r) + + result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + expected = run_opt_pass(expected(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) + + +def test_branch_merge(): + """Test composite function is correctly produced from branching graph. + + We would expect the pattern `make_add_sub_mul_pattern` to be merged + into a single op `add_sub_mul`. + + a b a b + \/ \/ + add sub a b + \ / \/ + \ / add_sub_mul + mul c | + / \ \ | + c / c | ====> add_sub_mul + \/ \/ | + add sub | + \ / ReLu + \ / + mul + | + | + ReLu + """ + + pattern_table = { + "add_sub_mul": make_add_sub_mul_pattern() + } + + def before(): + a = relay.var('a', shape=(10, 10)) + b = relay.var('b', shape=(10, 10)) + c = relay.var('c', shape=(10, 10)) + add_node = relay.add(a, b) + sub_node = relay.subtract(a, b) + mul_node = relay.multiply(add_node, sub_node) + add_node_2 = relay.add(c, mul_node) + sub_node_2 = relay.subtract(c, mul_node) + mul_node_2 = relay.multiply(add_node_2, sub_node_2) + r = relay.nn.relu(mul_node_2) + return relay.Function([a, b, c], r) + + def expected(): + a = relay.var('a', shape=(10, 10)) + b = relay.var('b', shape=(10, 10)) + c = relay.var('c', shape=(10, 10)) + + # add_sub_mul function + in_1 = relay.var('in_1', shape=(10, 10)) + in_2 = relay.var('in_2', shape=(10, 10)) + add_node = relay.add(in_1, in_2) + sub_node = relay.subtract(in_1, in_2) + mul_node = relay.multiply(add_node, sub_node) + add_sub_mul = relay.Function([in_1, in_2], mul_node) + + # merged function + add_sub_mul_1 = relay.Call(add_sub_mul, [a, b]) + add_sub_mul_2 = relay.Call(add_sub_mul, [c, add_sub_mul_1]) + r = relay.nn.relu(add_sub_mul_2) + return relay.Function([a, b, c], r) + + result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + expected = run_opt_pass(expected(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) From c6967ea8f8d7c1e88953342d01c69012af851c2a Mon Sep 17 00:00:00 2001 From: Luke Hutton Date: Wed, 22 Jan 2020 14:13:24 +0000 Subject: [PATCH 03/23] Merge composite additional test Change-Id: I9bc7d6053c575e9468ac5abc31214c6ad8507e46 --- .../python/relay/test_pass_merge_composite.py | 168 +++++++++++++++++- 1 file changed, 163 insertions(+), 5 deletions(-) diff --git a/tests/python/relay/test_pass_merge_composite.py b/tests/python/relay/test_pass_merge_composite.py index e44b18779f42..77b8f27902d1 100644 --- a/tests/python/relay/test_pass_merge_composite.py +++ b/tests/python/relay/test_pass_merge_composite.py @@ -18,6 +18,48 @@ from tvm import relay from tvm.relay.testing import run_opt_pass +""" +The merge composite pass is designed to merge multiple relay operators, that +match a given pattern, and combine them into a single relay function. + +For example suppose we have the graph: + + conv2d + | (merge composite pass) + bias_add ====> conv2d_bias_relu + | (our target) + relu + +Our Relay IR before the pass: + fn (%data: Tensor[(1, 512, 28, 28), float32], %kernel: Tensor[(256, 512, 1, 1), float32], + %bias: Tensor[(256), float32]) -> Tensor[(1, 256, 28, 28), float32] { + %0 = nn.conv2d(%data, %kernel, kernel_size=[1, 1]) + /* ty=Tensor[(1, 256, 28, 28), float32] */; + %1 = nn.bias_add(%0, %bias) /* ty=Tensor[(1, 256, 28, 28), float32] */; + nn.relu(%1) /* ty=Tensor[(1, 256, 28, 28), float32] */ + } + +Our Relay IR after the pass: + fn (%data: Tensor[(1, 512, 28, 28), float32], %kernel: Tensor[(256, 512, 1, 1), float32], + %bias: Tensor[(256), float32]) -> Tensor[(1, 256, 28, 28), float32] { + %2 = fn (%x: Tensor[(1, 512, 28, 28), float32], %y: Tensor[(256, 512, 1, 1), float32], + %z: Tensor[(256), float32], Primitive=1, Composite="conv2d_bias_relu") -> + Tensor[(1, 256, 28, 28), float32] { + %0 = nn.conv2d(%x, %y, kernel_size=[1, 1]) /* ty=Tensor[(1, 256, 28, 28), float32] */; + %1 = nn.bias_add(%0, %z) /* ty=Tensor[(1, 256, 28, 28), float32] */; + nn.relu(%1) /* ty=Tensor[(1, 256, 28, 28), float32] */ + }; + %2(%data, %kernel, %bias) /* ty=Tensor[(1, 256, 28, 28), float32] */ + } + +As you can see in the second relay example, the pattern we specified has been wrapped +in a function. The function is then called, producing the same result as the first relay +example. + +One convenient use for this pass is to offload multiple operators to a single external +codegen function. +""" + def make_add_sub_mul_pattern(): """Create a pattern to match the following graph. @@ -40,7 +82,7 @@ def make_add_relu_pattern(): add | - ReLu + relu """ x = relay.var('x') y = relay.var('y') @@ -49,6 +91,24 @@ def make_add_relu_pattern(): return r +def make_conv_bias_relu_pattern(): + """Create a pattern to match the following graph. + + conv2d + | + bias_add + | + relu + """ + x = relay.var('x') + y = relay.var('y') + z = relay.var('z') + conv_node = relay.nn.conv2d(x, y) + bias_node = relay.nn.bias_add(conv_node, z) + r = relay.nn.relu(bias_node) + return r + + def test_simple_merge(): """Test composite function is correctly produced from simple graph. @@ -59,11 +119,11 @@ def test_simple_merge(): \ / a b add ====> \ / | add_relu - ReLu + relu """ pattern_table = { - "add_sub_mul": make_add_relu_pattern() + "add_relu": make_add_relu_pattern() } def before(): @@ -89,6 +149,7 @@ def expected(): return relay.Function([a, b], r) result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) expected = run_opt_pass(expected(), relay.transform.InferType()) assert relay.analysis.alpha_equal(result, expected) @@ -109,12 +170,12 @@ def test_branch_merge(): c / c | ====> add_sub_mul \/ \/ | add sub | - \ / ReLu + \ / relu \ / mul | | - ReLu + relu """ pattern_table = { @@ -154,5 +215,102 @@ def expected(): return relay.Function([a, b, c], r) result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) + expected = run_opt_pass(expected(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) + + +def test_multiple_patterns(): + """Test different patterns are merged correctly in the graph. + + We would expect the pattern `make_conv_bias_relu_pattern` to be merged + into a single op `conv_bias_relu`. We would also expect `make_add_relu_pattern` + to be merged into a single op `add_relu`. + + data kernel + \ / + \ / + conv2d data kernel bias + | \ | / + | bias conv2d_bias_relu + | / | + bias_add ====> | a + | | / + relu a add_relu + \ / | + add | b + | | / + relu b mul + | / + mul + """ + pattern_table = { + "conv2d_bias_relu": make_conv_bias_relu_pattern(), + "add_relu": make_add_relu_pattern() + } + + def before(): + data = relay.var('data', shape=(1, 512, 28, 28)) + kernel = relay.var('kernel', shape=(256, 512, 1, 1)) + bias = relay.var('bias', shape=(256,)) + a = relay.var('a', shape=(1, 256, 28, 28)) + b = relay.var('b', shape=(1, 256, 28, 28)) + + conv_node = relay.nn.conv2d(data, + kernel, + kernel_size=(1, 1), + padding=(0, 0), + strides=(1, 1)) + + bias_node = relay.nn.bias_add(conv_node, bias) + relu_node = relay.nn.relu(bias_node) + add_node = relay.add(relu_node, a) + relu_node_2 = relay.nn.relu(add_node) + r = relay.multiply(relu_node_2, b) + return relay.Function([data, kernel, bias], r) + + def expected(): + data = relay.var('data', shape=(1, 512, 28, 28)) + kernel = relay.var('kernel', shape=(256, 512, 1, 1)) + bias = relay.var('bias', shape=(256,)) + a = relay.var('a', shape=(1, 256, 28, 28)) + b = relay.var('b', shape=(1, 256, 28, 28)) + + # conv_bias_relu function + in_1 = relay.var('in_1', shape=(1, 512, 28, 28)) + in_2 = relay.var('in_2', shape=(256, 512, 1, 1)) + in_3 = relay.var('in_3', shape=(256,)) + + conv_node = relay.nn.conv2d(in_1, + in_2, + kernel_size=(1, 1), + padding=(0, 0), + strides=(1, 1)) + + bias_node = relay.nn.bias_add(conv_node, in_3) + r = relay.nn.relu(bias_node) + conv_bias_add_relu = relay.Function([in_1, in_2, in_3], r) + + # add_relu function + in_4 = relay.var('in_4', shape=(1, 256, 28, 28)) + in_5 = relay.var('in_5', shape=(1, 256, 28, 28)) + add_node = relay.add(in_4, in_5) + r = relay.nn.relu(add_node) + add_relu = relay.Function([in_4, in_5], r) + + # merged function + conv_bias_add_relu_1 = relay.Call(conv_bias_add_relu, [data, kernel, bias]) + add_relu_1 = relay.Call(add_relu, [conv_bias_add_relu_1, a]) + r = relay.multiply(add_relu_1, b) + return relay.Function([data, kernel, bias, a, b], r) + + result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) expected = run_opt_pass(expected(), relay.transform.InferType()) assert relay.analysis.alpha_equal(result, expected) + + +if __name__ == "__main__": + test_simple_merge() + test_branch_merge() + test_multiple_patterns() From 19b5e77aac4fbbd5650cba3c44940c3b82c3f418 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 28 Jan 2020 17:27:31 +0000 Subject: [PATCH 04/23] Support priority order in merge_composite The order in which the patterns are matched was currently random as an unordered_map was used to store the pattern table. This uses arrays instead so that a distinct priority order of matching can be defined. Additional tests have also been added to verify this behaviour. Change-Id: Ief347df4262639138d5d9d7c8cee7ef233af7b56 --- python/tvm/relay/transform.py | 16 +- src/relay/pass/merge_composite.cc | 65 ++++---- .../python/relay/test_pass_merge_composite.py | 143 ++++++++++++++++-- 3 files changed, 182 insertions(+), 42 deletions(-) diff --git a/python/tvm/relay/transform.py b/python/tvm/relay/transform.py index d76727732842..cfca4a6ed3b2 100644 --- a/python/tvm/relay/transform.py +++ b/python/tvm/relay/transform.py @@ -513,13 +513,15 @@ def Legalize(legalize_map_attr_name="FTVMLegalize"): return _transform.Legalize(legalize_map_attr_name) -def MergeComposite(compiler): +def MergeComposite(pattern_table): """Merge multiple operators into a single composite relay function. Parameters ---------- - compiler : str - The compiler used for codegen. + pattern_table : list(tuple) + A list of (pattern_name, pattern) tuples. + The order of the patterns in the list will determine the order + of priority in which they are matched. Returns ------- @@ -527,7 +529,13 @@ def MergeComposite(compiler): The registered pass that merges operators into a single composite relay function. """ - return _transform.MergeComposite(compiler) + pattern_names = [] + patterns = [] + for pattern_name, pattern in pattern_table: + pattern_names.append(pattern_name) + patterns.append(pattern) + + return _transform.MergeComposite(pattern_names, patterns) def RewriteAnnotatedOps(fallback_device): diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 81d50b91a0b7..d4e445619190 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -36,8 +36,8 @@ namespace merge_composite { class MergeCompositeWrapper : public ExprMutator { public: - explicit MergeCompositeWrapper(const tvm::Map& pattern_map) - : pattern_map_(pattern_map) {} + explicit MergeCompositeWrapper(const std::string& pattern_name, const Expr& pattern) + : pattern_name_(pattern_name), pattern_(pattern) {} bool MatchPattern(const Call& pattern, const Call& root) { if (!pattern->op->IsInstance() || !root->op->IsInstance()) @@ -135,49 +135,58 @@ class MergeCompositeWrapper : public ExprMutator { Op op = Downcast(call->op); CHECK(op.defined()); - for (const auto& x : pattern_map_) { - Call pattern = Downcast(x.second); - if (Downcast(pattern->op)->name != op->name) - continue; - - if (MatchPattern(pattern, call)) { - Map> args_map; - auto extract = ExtractPattern(pattern, call, &args_map); - auto free_vars = FreeVars(extract); - Function new_func = FunctionNode::make(free_vars, extract, - call->checked_type_, {}, Attrs()); - new_func = FunctionSetAttr(new_func, attr::kComposite, - tir::StringImmNode::make(x.first)); - new_func = FunctionSetAttr(new_func, attr::kPrimitive, - tvm::Integer(1)); - Array args; - for (const auto& free_var : free_vars) { - args.push_back(args_map[free_var->name_hint()][1]); - } - auto new_call = CallNode::make(new_func, args); - return std::move(new_call); + Call pattern = Downcast(pattern_); + if (Downcast(pattern->op)->name != op->name) + return std::move(call); + + if (MatchPattern(pattern, call)) { + Map> args_map; + auto extract = ExtractPattern(pattern, call, &args_map); + auto free_vars = FreeVars(extract); + Function new_func = FunctionNode::make(free_vars, extract, + call->checked_type_, {}, Attrs()); + new_func = FunctionSetAttr(new_func, attr::kComposite, + tir::StringImmNode::make(pattern_name_)); + new_func = FunctionSetAttr(new_func, attr::kPrimitive, + tvm::Integer(1)); + Array args; + for (const auto& free_var : free_vars) { + args.push_back(args_map[free_var->name_hint()][1]); } + auto new_call = CallNode::make(new_func, args); + return std::move(new_call); } return std::move(call); } private: - tvm::Map pattern_map_; + std::string pattern_name_; + Expr pattern_; }; -Expr MergeComposite(const Expr& expr, const tvm::Map& pattern) { - return MergeCompositeWrapper(pattern).Mutate(expr); +Expr MergeComposite(const Expr& expr, + const Array& pattern_names, const Array& patterns) { + CHECK(pattern_names.size() == patterns.size()); + Expr merged_expr = expr; + for (size_t i = 0; i < patterns.size(); i++) { + std::string pattern_name = pattern_names[i]->value; + Expr pattern = patterns[i]; + merged_expr = MergeCompositeWrapper(pattern_name, pattern).Mutate(merged_expr); + } + return merged_expr; } } // namespace merge_composite namespace transform { -Pass MergeComposite(const tvm::Map& pattern) { +Pass MergeComposite(const tvm::Array& pattern_names, + const tvm::Array& patterns) { runtime::TypedPackedFunc pass_func = [=](Function f, IRModule m, PassContext pc) { - return Downcast(relay::merge_composite::MergeComposite(f, pattern)); + return Downcast( + relay::merge_composite::MergeComposite(f, pattern_names, patterns)); }; auto func_pass = CreateFunctionPass(pass_func, 0, "MergeComposite", {}); return func_pass; diff --git a/tests/python/relay/test_pass_merge_composite.py b/tests/python/relay/test_pass_merge_composite.py index 77b8f27902d1..53e428c04c17 100644 --- a/tests/python/relay/test_pass_merge_composite.py +++ b/tests/python/relay/test_pass_merge_composite.py @@ -15,6 +15,7 @@ # specific language governing permissions and limitations # under the License. """Unit tests for merge composite.""" +from tvm import expr from tvm import relay from tvm.relay.testing import run_opt_pass @@ -122,9 +123,9 @@ def test_simple_merge(): relu """ - pattern_table = { - "add_relu": make_add_relu_pattern() - } + pattern_table = [ + ("add_relu", make_add_relu_pattern()) + ] def before(): a = relay.var('a', shape=(10, 10)) @@ -178,9 +179,9 @@ def test_branch_merge(): relu """ - pattern_table = { - "add_sub_mul": make_add_sub_mul_pattern() - } + pattern_table = [ + ("add_sub_mul", make_add_sub_mul_pattern()) + ] def before(): a = relay.var('a', shape=(10, 10)) @@ -244,10 +245,10 @@ def test_multiple_patterns(): | / mul """ - pattern_table = { - "conv2d_bias_relu": make_conv_bias_relu_pattern(), - "add_relu": make_add_relu_pattern() - } + pattern_table = [ + ("conv2d_bias_relu", make_conv_bias_relu_pattern()), + ("add_relu", make_add_relu_pattern()) + ] def before(): data = relay.var('data', shape=(1, 512, 28, 28)) @@ -310,7 +311,129 @@ def expected(): assert relay.analysis.alpha_equal(result, expected) +def test_merge_order(): + """Test that patterns are merged in the order they exist in the pattern table. + + There can be cases where one pattern is a subgraph of another, in which case + it is not clear which match should take priority. The priority should come + from the order in which the patterns are declared in the pattern table. The + first patterns will be merged with highest priority and the last with lowest. + + A: B: C: + add add abs + | | | + abs abs relu + | + relu + + """ + + def pattern_A(): + x = relay.var('x') + y = relay.var('y') + out = relay.add(x, y) + out = relay.abs(out) + out = relay.nn.relu(out) + return out + + def pattern_B(): + x = relay.var('x') + y = relay.var('y') + out = relay.add(x, y) + out = relay.abs(out) + return out + + def pattern_C(): + x = relay.var('x') + out = relay.abs(x) + out = relay.nn.relu(x) + return out + + def before(): + input_1 = relay.var('input_1', shape=(10, 10)) + input_2 = relay.var('input_2', shape=(10, 10)) + out = relay.add(input_1, input_2) + out = relay.abs(out) + out = relay.nn.relu(out) + return relay.Function([input_1, input_2], out) + + def after_A_priority(): + input_1 = relay.var('input_1', shape=(10, 10)) + input_2 = relay.var('input_2', shape=(10, 10)) + x = relay.var('x') + y = relay.var('y') + out = relay.add(x, y) + out = relay.abs(out) + out = relay.nn.relu(out) + merged_func = relay.Function([x, y], out) + merged_func = merged_func.set_attribute('Primitive', expr.IntImm('int32', 1)) + merged_func = merged_func.set_attribute('Composite', expr.StringImm('A')) + ret = relay.Call(merged_func, [input_1, input_2]) + return relay.Function([input_1, input_2], ret) + + def after_B_priority(): + input_1 = relay.var('input_1', shape=(10, 10)) + input_2 = relay.var('input_2', shape=(10, 10)) + x = relay.var('x') + y = relay.var('y') + out = relay.add(x, y) + out = relay.abs(out) + merged_func = relay.Function([x, y], out) + merged_func = merged_func.set_attribute('Primitive', expr.IntImm('int32', 1)) + merged_func = merged_func.set_attribute('Composite', expr.StringImm('B')) + merged_call = relay.Call(merged_func, [input_1, input_2]) + ret = relay.nn.relu(merged_call) + return relay.Function([input_1, input_2], ret) + + def after_C_priority(): + input_1 = relay.var('input_1', shape=(10, 10)) + input_2 = relay.var('input_2', shape=(10, 10)) + add = relay.add(input_1, input_2) + x = relay.var('x') + out = relay.abs(x) + out = relay.nn.relu(out) + merged_func = relay.Function([x], out) + merged_func = merged_func.set_attribute('Primitive', expr.IntImm('int32', 1)) + merged_func = merged_func.set_attribute('Composite', expr.StringImm('C')) + ret = relay.Call(merged_func, [add]) + return relay.Function([input_1, input_2], ret) + + # check A highest priority + pattern_table = [ + ("A", pattern_A()), + ("B", pattern_B()), + ("C", pattern_C()), + ] + result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) + expected = run_opt_pass(after_A_priority(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) + + # check B highest priority + pattern_table = [ + ("B", pattern_A()), + ("C", pattern_B()), + ("A", pattern_C()), + ] + result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) + expected = run_opt_pass(after_A_priority(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) + + # check C highest priority + pattern_table = [ + ("C", pattern_A()), + ("A", pattern_B()), + ("B", pattern_C()), + ] + result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) + expected = run_opt_pass(after_A_priority(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) + + if __name__ == "__main__": test_simple_merge() test_branch_merge() test_multiple_patterns() + test_merge_order() From 66f5ec9bb4ac08e7a16276e7e3da29d90c00441f Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 10:34:45 +0000 Subject: [PATCH 05/23] Improved merge composite docs Change-Id: Ie3a72045ecc3f13ad3c302fbdf192b7296a306a8 --- src/relay/pass/merge_composite.cc | 13 +++++++++++-- 1 file changed, 11 insertions(+), 2 deletions(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index d4e445619190..303fe6169e8a 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -20,7 +20,9 @@ /*! * \file src/relay/pass/merge_composite.cc * \brief Merges expressions matching patterns into functions marked - * as 'composite'. + * as 'composite'. This is primarily intended to be used alongside the + * external codegen infrastructure to support the case where multiple + * Relay operators map to a single external operator. */ #include @@ -38,7 +40,12 @@ class MergeCompositeWrapper : public ExprMutator { public: explicit MergeCompositeWrapper(const std::string& pattern_name, const Expr& pattern) : pattern_name_(pattern_name), pattern_(pattern) {} - + /*! + * \brief Determine if a pattern and a subgraph have the same call structure. + * \param pattern the root call node of the pattern. + * \param root the root call node of the subgraph. + * \return Whether they match. + */ bool MatchPattern(const Call& pattern, const Call& root) { if (!pattern->op->IsInstance() || !root->op->IsInstance()) return false; @@ -118,6 +125,7 @@ class MergeCompositeWrapper : public ExprMutator { Function func = Downcast(call->op); CHECK(func.defined()); const auto name_node = FunctionGetAttr(func, attr::kComposite).as(); + // don't step into existing composite functions if (name_node->value != "") { tvm::Array new_args; for (const auto& arg : call->args) { @@ -169,6 +177,7 @@ Expr MergeComposite(const Expr& expr, const Array& pattern_names, const Array& patterns) { CHECK(pattern_names.size() == patterns.size()); Expr merged_expr = expr; + // merge the patterns one-by-one in order for (size_t i = 0; i < patterns.size(); i++) { std::string pattern_name = pattern_names[i]->value; Expr pattern = patterns[i]; From dc23021dd35491b1fabb03c76fb1f0b42211d691 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 10:42:39 +0000 Subject: [PATCH 06/23] Removed unused variable Change-Id: I7814d5fde368ffaf1b3d6d806060c774c7720364 --- src/relay/pass/merge_composite.cc | 1 - 1 file changed, 1 deletion(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 303fe6169e8a..cd6f9408fb9f 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -85,7 +85,6 @@ class MergeCompositeWrapper : public ExprMutator { Expr ExtractPattern(const Call& pattern, const Call& root, Map>* var_map) { - Expr expr; Expr empty_expr; if (!pattern->op->IsInstance() || !root->op->IsInstance()) return empty_expr; From a00a29af6d7b7150be689d332892531eb2de79ab Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 10:47:38 +0000 Subject: [PATCH 07/23] Remove unnecessary op check Change-Id: I38e78d2acd5b86cb8e837be72ff9d72cd10bcf33 --- src/relay/pass/merge_composite.cc | 7 ++----- 1 file changed, 2 insertions(+), 5 deletions(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index cd6f9408fb9f..024c8ab622da 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -140,12 +140,9 @@ class MergeCompositeWrapper : public ExprMutator { if (!call->op->IsInstance()) return std::move(call); - Op op = Downcast(call->op); - CHECK(op.defined()); + // only call patterns are currently supported Call pattern = Downcast(pattern_); - if (Downcast(pattern->op)->name != op->name) - return std::move(call); - + CHECK(pattern.defined()); if (MatchPattern(pattern, call)) { Map> args_map; auto extract = ExtractPattern(pattern, call, &args_map); From c2db7de1579cc4984682ddc4389c27c1f58402b1 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 10:56:35 +0000 Subject: [PATCH 08/23] Improve styling on composite function creation Change-Id: I37add1c3134e0b5d5085fe1eb9daf8e06890fa8c --- src/relay/pass/merge_composite.cc | 12 +++++------- 1 file changed, 5 insertions(+), 7 deletions(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 024c8ab622da..2d843d3aa8e6 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -147,17 +147,15 @@ class MergeCompositeWrapper : public ExprMutator { Map> args_map; auto extract = ExtractPattern(pattern, call, &args_map); auto free_vars = FreeVars(extract); - Function new_func = FunctionNode::make(free_vars, extract, - call->checked_type_, {}, Attrs()); - new_func = FunctionSetAttr(new_func, attr::kComposite, - tir::StringImmNode::make(pattern_name_)); - new_func = FunctionSetAttr(new_func, attr::kPrimitive, - tvm::Integer(1)); + // make the composite function + auto f = FunctionNode::make(free_vars, extract, call->checked_type_, {}, Attrs()); + f = FunctionSetAttr(f, attr::kComposite, tir::StringImmNode::make(pattern_name_)); + f = FunctionSetAttr(f, attr::kPrimitive, tvm::Integer(1)); Array args; for (const auto& free_var : free_vars) { args.push_back(args_map[free_var->name_hint()][1]); } - auto new_call = CallNode::make(new_func, args); + auto new_call = CallNode::make(f, args); return std::move(new_call); } From 507db6fb01e42bfd0ce3213c50cded0830397ca2 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 10:58:32 +0000 Subject: [PATCH 09/23] Comment reword Change-Id: Ie05872dcbbe0c3e1190b0597083b9a64e6b66c66 --- src/relay/pass/merge_composite.cc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 2d843d3aa8e6..639936991417 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -140,7 +140,7 @@ class MergeCompositeWrapper : public ExprMutator { if (!call->op->IsInstance()) return std::move(call); - // only call patterns are currently supported + // only call patterns are supported Call pattern = Downcast(pattern_); CHECK(pattern.defined()); if (MatchPattern(pattern, call)) { From 3ab884a97946ae79b51f7d31b5d79a9aa5944189 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 11:16:44 +0000 Subject: [PATCH 10/23] Stylistic changes to avoid std::move Change-Id: I43a93995bbf10530399900c992aa99dd4ae4575f --- src/relay/pass/merge_composite.cc | 37 ++++++++++++++++--------------- 1 file changed, 19 insertions(+), 18 deletions(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 639936991417..943ee14c9bfa 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -85,37 +85,38 @@ class MergeCompositeWrapper : public ExprMutator { Expr ExtractPattern(const Call& pattern, const Call& root, Map>* var_map) { - Expr empty_expr; if (!pattern->op->IsInstance() || !root->op->IsInstance()) - return empty_expr; + return Expr(); if (pattern->op.as()->name != root->op.as()->name) - return empty_expr; + return Expr(); if (pattern->args.size() != root->args.size()) - return empty_expr; + return Expr(); unsigned int i = 0; Array new_args; for (const auto& arg : pattern->args) { + Expr new_arg; if (arg->IsInstance()) { - new_args.push_back(ExtractPattern(Downcast(arg), - Downcast(root->args[i]), - var_map)); + new_arg = ExtractPattern(Downcast(arg), + Downcast(root->args[i]), + var_map); + } else if (arg->IsInstance()) { + new_arg = ExtractPattern(Downcast(arg), + root->args[i], + var_map); + } else if (arg->IsInstance()) { + new_arg = ExtractPattern(Downcast(arg), + root->args[i], + var_map); } - if (arg->IsInstance()) { - new_args.push_back(ExtractPattern(Downcast(arg), - root->args[i], - var_map)); - } - if (arg->IsInstance()) { - new_args.push_back(ExtractPattern(Downcast(arg), - root->args[i], - var_map)); + if (!new_arg.defined()) { + return Expr(); } + new_args.push_back(new_arg); i++; } - auto new_call = CallNode::make(root->op, new_args, root->attrs); - return std::move(new_call); + return CallNode::make(root->op, new_args, root->attrs); } Expr VisitExpr_(const CallNode* cn) { From 31b04b51cdaafb3ec00fca7deb41aa1d7097a2e2 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 11:25:11 +0000 Subject: [PATCH 11/23] Relax a check in ExtractPattern Change-Id: I0faef77a66c55f83f09e6e47c561ffaea63dedfa --- src/relay/pass/merge_composite.cc | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 943ee14c9bfa..71ecc13589b3 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -85,12 +85,11 @@ class MergeCompositeWrapper : public ExprMutator { Expr ExtractPattern(const Call& pattern, const Call& root, Map>* var_map) { + // check to make sure both calls are to operators (not functions) if (!pattern->op->IsInstance() || !root->op->IsInstance()) return Expr(); if (pattern->op.as()->name != root->op.as()->name) return Expr(); - if (pattern->args.size() != root->args.size()) - return Expr(); unsigned int i = 0; Array new_args; From 73cd0263d2fcdaab5400366d18f644d82036efd9 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 11:25:42 +0000 Subject: [PATCH 12/23] Remove new line Change-Id: Ifdd02c12087a7e1a0a9b54825669bc0de8f13c3d --- src/relay/pass/merge_composite.cc | 1 - 1 file changed, 1 deletion(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 71ecc13589b3..02a0b0f172f5 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -114,7 +114,6 @@ class MergeCompositeWrapper : public ExprMutator { new_args.push_back(new_arg); i++; } - return CallNode::make(root->op, new_args, root->attrs); } From f76389836c5c3910f76c93bb0963e8d4b13dbacf Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 11:34:53 +0000 Subject: [PATCH 13/23] Removed MatchPattern from MergeComposite This is not necessary now that ExtractPattern can fulfill the same purpose. Change-Id: I14dc020afa8e50f2df4c0a2efb88a011987f8196 --- src/relay/pass/merge_composite.cc | 32 +++---------------------------- 1 file changed, 3 insertions(+), 29 deletions(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 02a0b0f172f5..c0fe6a56e351 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -40,32 +40,6 @@ class MergeCompositeWrapper : public ExprMutator { public: explicit MergeCompositeWrapper(const std::string& pattern_name, const Expr& pattern) : pattern_name_(pattern_name), pattern_(pattern) {} - /*! - * \brief Determine if a pattern and a subgraph have the same call structure. - * \param pattern the root call node of the pattern. - * \param root the root call node of the subgraph. - * \return Whether they match. - */ - bool MatchPattern(const Call& pattern, const Call& root) { - if (!pattern->op->IsInstance() || !root->op->IsInstance()) - return false; - if (pattern->op.as()->name != root->op.as()->name) - return false; - if (pattern->args.size() != root->args.size()) - return false; - - unsigned int i = 0; - for (const auto& arg : pattern->args) { - if (arg->IsInstance()) { - if (!root->args[i]->IsInstance()) - return false; - if (!MatchPattern(Downcast(arg), Downcast(root->args[i]))) - return false; - } - i++; - } - return true; - } Expr ExtractPattern(const Var& pattern, const Expr& root, Map>* var_map) { @@ -142,9 +116,9 @@ class MergeCompositeWrapper : public ExprMutator { // only call patterns are supported Call pattern = Downcast(pattern_); CHECK(pattern.defined()); - if (MatchPattern(pattern, call)) { - Map> args_map; - auto extract = ExtractPattern(pattern, call, &args_map); + Map> args_map; + auto extract = ExtractPattern(pattern, call, &args_map); + if (extract.defined()) { auto free_vars = FreeVars(extract); // make the composite function auto f = FunctionNode::make(free_vars, extract, call->checked_type_, {}, Attrs()); From 25f67e111a6f652d6790fcead2aee0d6bdd705cb Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 11:35:52 +0000 Subject: [PATCH 14/23] Removed a new line Change-Id: I8b50f0c9069aa1bcaccbe68eb421031f01a64842 --- src/relay/pass/merge_composite.cc | 1 - 1 file changed, 1 deletion(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index c0fe6a56e351..385eac118c57 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -131,7 +131,6 @@ class MergeCompositeWrapper : public ExprMutator { auto new_call = CallNode::make(f, args); return std::move(new_call); } - return std::move(call); } From 49c8d8f6d6cd3cbb7b48183030445b1e381b9bb2 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 12:06:20 +0000 Subject: [PATCH 15/23] Improved docs for merge composite Change-Id: Ib1959a35c856e7ea5639de2e4ef314a54f44caf5 --- src/relay/pass/merge_composite.cc | 25 +++++++++++++++++++++++++ 1 file changed, 25 insertions(+) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 385eac118c57..da48d7b2fa3a 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -44,10 +44,13 @@ class MergeCompositeWrapper : public ExprMutator { Expr ExtractPattern(const Var& pattern, const Expr& root, Map>* var_map) { if (var_map->find(pattern->name_hint()) == var_map->end()) { + // if we haven't encountered this var yet, make a new free var and associate + // it with the value at 'root' auto free_var = VarNode::make(pattern->name_hint(), Type()); var_map->Set(pattern->name_hint(), Array({free_var, root})); return std::move(free_var); } else { + // if we have encountered this var already, return the free var that was created return (*var_map)[pattern->name_hint()][0]; } } @@ -57,6 +60,19 @@ class MergeCompositeWrapper : public ExprMutator { return root; } + /* How does this work? + * + * A pattern consists of Relay expression containing only operator call nodes, constants + * and free variables. The free variables indicate where the pattern can 'attach' in your + * graph. This function takes the final call node of the pattern and the call node currently + * being traversed in the Relay graph. It traverses through the pattern in lockstep with call node + * from the graph (referred to as the 'root' node here) to check they're identical. If at any point + * they differ, an empty expression is returned to signify the extract failed. If a free var is + * reached in the pattern, the corresponding value in the root is associated with the name of the + * free var (via the var_map) so that when we construct the composite function, the inputs match + * up correctly with the rest of the graph. The return value of this function when successful is + * a new Relay expression ready to be wrapped into a composite function. + */ Expr ExtractPattern(const Call& pattern, const Call& root, Map>* var_map) { // check to make sure both calls are to operators (not functions) @@ -70,14 +86,19 @@ class MergeCompositeWrapper : public ExprMutator { for (const auto& arg : pattern->args) { Expr new_arg; if (arg->IsInstance()) { + // if it's a call node, recursively call this function new_arg = ExtractPattern(Downcast(arg), Downcast(root->args[i]), var_map); } else if (arg->IsInstance()) { + // if there's a var in the pattern, it must be a free var + // so call the function to update the var_map new_arg = ExtractPattern(Downcast(arg), root->args[i], var_map); } else if (arg->IsInstance()) { + // if there's a constant, simply get the corresponding + // value of the constant from the root new_arg = ExtractPattern(Downcast(arg), root->args[i], var_map); @@ -124,6 +145,8 @@ class MergeCompositeWrapper : public ExprMutator { auto f = FunctionNode::make(free_vars, extract, call->checked_type_, {}, Attrs()); f = FunctionSetAttr(f, attr::kComposite, tir::StringImmNode::make(pattern_name_)); f = FunctionSetAttr(f, attr::kPrimitive, tvm::Integer(1)); + // find the expressions associated with the free vars using the args_map + // this tells us which expressions should be given as inputs to the composite function Array args; for (const auto& free_var : free_vars) { args.push_back(args_map[free_var->name_hint()][1]); @@ -135,7 +158,9 @@ class MergeCompositeWrapper : public ExprMutator { } private: + /*! \brief The name of the pattern to match */ std::string pattern_name_; + /*! \brief The pattern to match */ Expr pattern_; }; From 66c97d9ca9e29fef2a6316189fe4d6c4c29798c4 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 12:09:03 +0000 Subject: [PATCH 16/23] Fixed free vars in test Change-Id: I2b7f273db275964ec0e9820560663f0808adee79 --- tests/python/relay/test_pass_merge_composite.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/python/relay/test_pass_merge_composite.py b/tests/python/relay/test_pass_merge_composite.py index 53e428c04c17..169f3c2c097b 100644 --- a/tests/python/relay/test_pass_merge_composite.py +++ b/tests/python/relay/test_pass_merge_composite.py @@ -268,7 +268,7 @@ def before(): add_node = relay.add(relu_node, a) relu_node_2 = relay.nn.relu(add_node) r = relay.multiply(relu_node_2, b) - return relay.Function([data, kernel, bias], r) + return relay.Function([data, kernel, bias, a, b], r) def expected(): data = relay.var('data', shape=(1, 512, 28, 28)) From dff2ff6ff4c01e50b7d93adfc6c3b0d8585ff0d7 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Tue, 4 Feb 2020 12:15:33 +0000 Subject: [PATCH 17/23] Handle case where root arg might not be a call Change-Id: I4eeea3ce723d3ba337d110dcc690377daebe8626 --- src/relay/pass/merge_composite.cc | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index da48d7b2fa3a..b6d74c0a34b5 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -86,6 +86,10 @@ class MergeCompositeWrapper : public ExprMutator { for (const auto& arg : pattern->args) { Expr new_arg; if (arg->IsInstance()) { + // fail if the root argument is not also a call node + if (!root->args[i]->IsInstance()) { + return Expr(); + } // if it's a call node, recursively call this function new_arg = ExtractPattern(Downcast(arg), Downcast(root->args[i]), From 7a96fb9a578aa0607b9fe3c9ddfd98f07959595f Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Mon, 10 Feb 2020 12:18:48 +0000 Subject: [PATCH 18/23] Removed blank line Change-Id: I07f5392c0e95cfe3cfa5c333703cc6f82d6034fb --- src/relay/pass/merge_composite.cc | 1 - 1 file changed, 1 deletion(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index b6d74c0a34b5..25d3ef7c9942 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -35,7 +35,6 @@ namespace tvm { namespace relay { namespace merge_composite { - class MergeCompositeWrapper : public ExprMutator { public: explicit MergeCompositeWrapper(const std::string& pattern_name, const Expr& pattern) From 7f1c9b25a171184d640cde71a33d48bf823de139 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Mon, 10 Feb 2020 12:19:19 +0000 Subject: [PATCH 19/23] Change to CHECK_EQ Change-Id: I5c5d62d3cd57f72508b30b926f72091ae6f0d1cc --- src/relay/pass/merge_composite.cc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 25d3ef7c9942..e38ebc0366e2 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -169,7 +169,7 @@ class MergeCompositeWrapper : public ExprMutator { Expr MergeComposite(const Expr& expr, const Array& pattern_names, const Array& patterns) { - CHECK(pattern_names.size() == patterns.size()); + CHECK_EQ(pattern_names.size(), patterns.size()); Expr merged_expr = expr; // merge the patterns one-by-one in order for (size_t i = 0; i < patterns.size(); i++) { From 14c17bd4a1671b1ae50c09bc063125771c743a29 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Mon, 10 Feb 2020 12:22:01 +0000 Subject: [PATCH 20/23] Revised a conditional Change-Id: I23a7897ca15a7cd076db5039dc653a4b8c27e803 --- src/relay/pass/merge_composite.cc | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index e38ebc0366e2..e4f54c5f7a1c 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -122,7 +122,7 @@ class MergeCompositeWrapper : public ExprMutator { CHECK(func.defined()); const auto name_node = FunctionGetAttr(func, attr::kComposite).as(); // don't step into existing composite functions - if (name_node->value != "") { + if (name_node && name_node->value != "") { tvm::Array new_args; for (const auto& arg : call->args) { auto new_e = this->Mutate(arg); From 1fb31a16f753b15664986092d38cb018883aed0c Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Mon, 10 Feb 2020 13:45:23 +0000 Subject: [PATCH 21/23] Improved doc styling Change-Id: I377f0a1c1ac70f3b8d7584b0c49bddc8c6c134ef --- src/relay/pass/merge_composite.cc | 9 ++++++++- 1 file changed, 8 insertions(+), 1 deletion(-) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index e4f54c5f7a1c..551249590e18 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -59,7 +59,14 @@ class MergeCompositeWrapper : public ExprMutator { return root; } - /* How does this work? + /*! + * \brief Try and extract a given pattern from a graph as a subgraph. + * \param pattern The pattern to extract. + * \param root The graph to extract from. + * \param var_map A map between free vars in the subgraph and nodes in the graph. + * \return The extracted subgraph. + * + * \note How does this work? * * A pattern consists of Relay expression containing only operator call nodes, constants * and free variables. The free variables indicate where the pattern can 'attach' in your From 7d63fd16ef1f0e70725c6f6bacd96d6854779ed9 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Mon, 10 Feb 2020 15:27:02 +0000 Subject: [PATCH 22/23] Fail extraction if vars conflict Change-Id: I78e36d805e8ed6b55e61d490212a967c857554a4 --- src/relay/pass/merge_composite.cc | 7 +++++++ 1 file changed, 7 insertions(+) diff --git a/src/relay/pass/merge_composite.cc b/src/relay/pass/merge_composite.cc index 551249590e18..28bf8fa8c33a 100644 --- a/src/relay/pass/merge_composite.cc +++ b/src/relay/pass/merge_composite.cc @@ -50,6 +50,13 @@ class MergeCompositeWrapper : public ExprMutator { return std::move(free_var); } else { // if we have encountered this var already, return the free var that was created + auto vars = (*var_map)[pattern->name_hint()]; + auto free_var = vars[0]; + auto graph_expr = vars[1]; + // make sure to first check they both map to the same node in the graph + if (graph_expr != root) { + return Expr(); + } return (*var_map)[pattern->name_hint()][0]; } } From f156aab7954f65e8284b90eacb165a8648044812 Mon Sep 17 00:00:00 2001 From: Matthew Barrett Date: Mon, 10 Feb 2020 15:27:29 +0000 Subject: [PATCH 23/23] Added further merge composite tests Change-Id: Ib1d800409fca4c1834c7fe0cab5a26ab99a26820 --- .../python/relay/test_pass_merge_composite.py | 170 ++++++++++++++++++ 1 file changed, 170 insertions(+) diff --git a/tests/python/relay/test_pass_merge_composite.py b/tests/python/relay/test_pass_merge_composite.py index 169f3c2c097b..4f785d7c915e 100644 --- a/tests/python/relay/test_pass_merge_composite.py +++ b/tests/python/relay/test_pass_merge_composite.py @@ -432,8 +432,178 @@ def after_C_priority(): assert relay.analysis.alpha_equal(result, expected) +def test_parallel_merge(): + """Tests that parallel patterns relying on the same inputs are correctly merged. + + The test graph is difficult to draw out as ascii art. It is essentially two parallel + add-sub-mul units which both consume input_1 and input_2 with their results being multiplied + to give the output. We expect both parallel branches should get merged and both should still + consume the same input variables, input_1 and input_2.""" + + def before(): + input_1 = relay.var('input_1', shape=(10, 10)) + input_2 = relay.var('input_2', shape=(10, 10)) + branch_1_add = relay.add(input_1, input_2) + branch_1_sub = relay.subtract(input_1, input_2) + branch_1 = relay.multiply(branch_1_add, branch_1_sub) + branch_2_add = relay.add(input_1, input_2) + branch_2_sub = relay.subtract(input_1, input_2) + branch_2 = relay.multiply(branch_2_add, branch_2_sub) + out = relay.multiply(branch_1, branch_2) + return relay.Function([input_1, input_2], out) + + def after(): + input_1 = relay.var('input_1', shape=(10, 10)) + input_2 = relay.var('input_2', shape=(10, 10)) + x = relay.var('x') + y = relay.var('y') + branch_1 = relay.multiply(relay.add(x, y), relay.subtract(x, y)) + func_1 = relay.Function([x, y], branch_1) + call_1 = relay.Call(func_1, [input_1, input_2]) + x1 = relay.var('x1') + y1 = relay.var('y1') + branch_2 = relay.multiply(relay.add(x1, y1), relay.subtract(x1, y1)) + func_2 = relay.Function([x1, y1], branch_2) + call_2 = relay.Call(func_2, [input_1, input_2]) + out = relay.multiply(call_1, call_2) + return relay.Function([input_1, input_2], out) + + pattern_table = [ + ("add_sub_mul", make_add_sub_mul_pattern()) + ] + result = run_opt_pass(before(), relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) + expected = run_opt_pass(after(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) + + +def test_multiple_input_subgraphs(): + """Test the case when multiple input subgraphs feed into another subgraph. + + (1) (2) (3) (4) + add add add add + | | | | + relu relu relu relu + \ / \ / + \ / \ / + add sub + \ / + \ / + \ / + mul + + ----> When 1=3 and 2=4 (Case 'A') + + add_relu add_relu + \ / + \ / + add_sub_mul + + ----> When 1!=3 and 2!=4 (Case 'B') + + add_relu add_relu add_relu add_relu + \ / \ / + \ / \ / + add sub + \ / + -------- ----- + \ / + mul + + The difference in behaviour comes from the fact that add_sub_mul expects that the + inputs to add and sub are identical (the same two relay expressions). So when you + have 4 independent inputs, the pattern should not be merged. + """ + + def before(): + before_funcs = {} + inputs = [relay.var('input_' + str(i), shape=(10, 10)) for i in range(8)] + add_relu_1 = relay.add(inputs[0], inputs[1]) + add_relu_1 = relay.nn.relu(add_relu_1) + add_relu_2 = relay.add(inputs[2], inputs[3]) + add_relu_2 = relay.nn.relu(add_relu_2) + add_relu_3 = relay.add(inputs[4], inputs[5]) + add_relu_3 = relay.nn.relu(add_relu_3) + add_relu_4 = relay.add(inputs[6], inputs[7]) + add_relu_4 = relay.nn.relu(add_relu_4) + add = relay.add(add_relu_1, add_relu_2) + sub = relay.subtract(add_relu_3, add_relu_4) + out = relay.multiply(add, sub) + before_funcs['B'] = relay.Function(inputs, out) + sub = relay.subtract(add_relu_1, add_relu_2) + out = relay.multiply(add, sub) + before_funcs['A'] = relay.Function(inputs[:4], out) + return before_funcs + + def after_A(): + inputs = [relay.var('input_' + str(i), shape=(10, 10)) for i in range(4)] + x = relay.var('x') + y = relay.var('y') + add_relu_1 = relay.add(x, y) + add_relu_1 = relay.nn.relu(add_relu_1) + add_relu_1 = relay.Function([x, y], add_relu_1) + add_relu_1 = add_relu_1.set_attribute('Primitive', expr.IntImm('int32', 1)) + add_relu_1 = add_relu_1.set_attribute('Composite', expr.StringImm('add_relu')) + add_relu_call_1 = relay.Call(add_relu_1, [inputs[0], inputs[1]]) + x1 = relay.var('x1') + y1 = relay.var('y1') + add_relu_2 = relay.add(x1, y1) + add_relu_2 = relay.nn.relu(add_relu_2) + add_relu_2 = relay.Function([x1, y1], add_relu_2) + add_relu_2 = add_relu_2.set_attribute('Primitive', expr.IntImm('int32', 1)) + add_relu_2 = add_relu_2.set_attribute('Composite', expr.StringImm('add_relu')) + add_relu_call_2 = relay.Call(add_relu_2, [inputs[2], inputs[3]]) + x2 = relay.var('x2') + y2 = relay.var('y2') + add = relay.add(x2, y2) + sub = relay.subtract(x2, y2) + add_sub_mul = relay.multiply(add, sub) + add_sub_mul = relay.Function([x2, y2], add_sub_mul) + add_sub_mul = add_sub_mul.set_attribute('Primitive', expr.IntImm('int32', 1)) + add_sub_mul = add_sub_mul.set_attribute('Composite', expr.StringImm('add_sub_mul')) + add_sub_mul_call = relay.Call(add_sub_mul, [add_relu_call_1, add_relu_call_2]) + return relay.Function(inputs, add_sub_mul_call) + + def after_B(): + inputs = [relay.var('input_' + str(i), shape=(10, 10)) for i in range(8)] + add_relu_calls = [] + for i in range(4): + x = relay.var('x' + str(i)) + y = relay.var('x' + str(i)) + add_relu = relay.add(x, y) + add_relu = relay.nn.relu(add_relu) + add_relu = relay.Function([x, y], add_relu) + add_relu = add_relu.set_attribute('Primitive', expr.IntImm('int32', 1)) + add_relu = add_relu.set_attribute('Composite', expr.StringImm('add_relu')) + add_relu_call = relay.Call(add_relu, [inputs[i*2], inputs[i*2+1]]) + add_relu_calls.append(add_relu_call) + + add = relay.add(add_relu_calls[0], add_relu_calls[1]) + sub = relay.subtract(add_relu_calls[2], add_relu_calls[3]) + out = relay.multiply(add, sub) + return relay.Function(inputs, out) + + pattern_table = [ + ("add_sub_mul", make_add_sub_mul_pattern()), + ("add_relu", make_add_relu_pattern()) + ] + # check case 'A' + result = run_opt_pass(before()['A'], relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) + expected = run_opt_pass(after_A(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) + + # check case 'B' + result = run_opt_pass(before()['B'], relay.transform.MergeComposite(pattern_table)) + assert not relay.analysis.free_vars(result) + expected = run_opt_pass(after_B(), relay.transform.InferType()) + assert relay.analysis.alpha_equal(result, expected) + + if __name__ == "__main__": test_simple_merge() test_branch_merge() test_multiple_patterns() test_merge_order() + test_parallel_merge() + test_multiple_input_subgraphs() \ No newline at end of file