LayerNorm OP draft#1186
Conversation
|
ok |
lambda7xx
left a comment
There was a problem hiding this comment.
Reviewable status: 0 of 5 files reviewed, 17 unresolved discussions (waiting on @lambda7xx and @reyna-abhyankar)
lib/runtime/src/ops/layer_norm.cc line 0 at r1 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
(Reviewable was unable to map this GitHub inline comment thread to the right spot — sorry!)
Keep these
using Legion::Context; using Legion::PhysicalRegion; using Legion::Runtime; using Legion::Task;
Done.
lib/runtime/src/ops/layer_norm.cc line 0 at r1 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
(Reviewable was unable to map this GitHub inline comment thread to the right spot — sorry!)
@lambda7xx this is the original code that sets
effective_num_elementsandeffective_batch_size. Seems like we can get this frominput
Done.
lib/runtime/src/ops/layer_norm.cc line 0 at r1 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
Split the signature out of register task (see other merged PRs)
Done.
lib/runtime/src/ops/layer_norm.cc line 43 at r1 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
I think this is correct since gamma and beta are parameters
Done.
lib/runtime/src/ops/layer_norm.cc line 60 at r1 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
Based on original permissions, I think it's RW and for backward only the grads are RW
Done.
lib/runtime/src/ops/layer_norm.cc line 144 at r1 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
InputParallelTensorDesc const &input_shape,I think we should use
InputParallelTensorDescfor input shape always because it also has theIsTrainable
Done.
lib/runtime/src/ops/layer_norm.cc line 162 at r1 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
I think it's the same shape for both
Done.
lib/runtime/src/ops/layer_norm.cc line 195 at r1 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
fwd.add_weight_slot(GAMMA); fwd.add_weight_slot(BETA);
Done.
lib/runtime/src/ops/layer_norm.cc line 138 at r2 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
In the original code I don't think it's divided by
num_replicasto geteffective_batch_size. Justinputs->volume() / M
Done.
lib/runtime/src/ops/layer_norm.cc line 138 at r2 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
Done.
lib/runtime/src/ops/layer_norm.cc line 141 at r2 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
effective_num_elements = M; effective_batch_size = input.shape.get_volume() / M;
Done.
lib/runtime/src/ops/layer_norm.cc line 187 at r2 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
fwd_binding.bind(GAMMA, input_shape); fwd_binding.bind(BETA, input_shape);
Done.
lib/runtime/src/ops/layer_norm.cc line 217 at r4 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
OpTaskSignature bwd_signature<LAYERNORM_BWD_TASK_ID>() {
Done.
lib/runtime/src/ops/layer_norm.cc line 223 at r4 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
OpTaskSignature init_signature<LAYERNORM_INIT_TASK_ID>() {
Done.
lib/runtime/src/ops/layer_norm.cc line 236 at r4 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
register_task(LAYERNORM_INIT_TASK_ID, "LayerNorm init", init_signature<LAYERNORM_INIT_TASK_ID>(), init_task);
Done.
lib/runtime/src/ops/layer_norm.cc line 247 at r4 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
LAYERNORM_BWD_TASK_ID, "LayerNorm backward", bwd_signature<AYERNORM_BWD_TASK_ID>() , backward_task);
Done.
deps/fmt line 0 at r2 (raw file):
Previously, reyna-abhyankar (Reyna Abhyankar) wrote…
Use most updated submodules
Done.
Description of changes:
int64_t effective_batch_size, effective_num_elementsin LayerNormPerDeviceStategamma and betain forward and backwardRelated Issues:
Linked Issues:
Issues closed by this PR:
This change is