Merged
Conversation
Collaborator
Author
|
!build |
naoyam
reviewed
Jun 7, 2023
This currently fails at lowering due to infinite recursion in nvfuser::prove::lessEqual when trying to simplify index expressions for index hoisting.
jacobhinkle
commented
Jun 7, 2023
Collaborator
Author
|
Closing in favor of #511. |
jacobhinkle
commented
Sep 11, 2023
|
|
||
| const int64_t slice_offset = 4; | ||
| const std::vector<int64_t> shape({1024 * 1024}); | ||
| const std::vector<int64_t> shape({1024L * 1024L}); |
Collaborator
Author
There was a problem hiding this comment.
Silencing clang-tidy
jacobhinkle
commented
Sep 14, 2023
jacobhinkle
commented
Sep 14, 2023
jacobhinkle
commented
Sep 14, 2023
jacobhinkle
commented
Sep 14, 2023
Comment on lines
-898
to
+899
| std::cout << "Fusion IR after pre-segmenter optimization passes:" | ||
| << std::endl; | ||
| debug() << "Fusion IR after pre-segmenter optimization passes:" | ||
| << std::endl; |
Collaborator
Author
There was a problem hiding this comment.
Unrelated to this PR. Just found wrong ostream in this debug dump.
zasdfgbnm
reviewed
Sep 18, 2023
Collaborator
Author
|
!build |
Collaborator
Author
|
!build |
zasdfgbnm
reviewed
Sep 26, 2023
zasdfgbnm
reviewed
Sep 26, 2023
zasdfgbnm
reviewed
Sep 26, 2023
test/test_resize.cpp
Outdated
|
|
||
| // Test slice with a variety of constant ranges | ||
| TEST_F(NVFuserTest, FusionResizeSliceConstantShmoo_CUDA) { | ||
| for (auto [start, stop] : std::vector<std::pair<int64_t, int64_t>>( |
Collaborator
There was a problem hiding this comment.
Should we use the same set of slices as FusionResizeSliceInputShmoo_CUDA?
Collaborator
Author
There was a problem hiding this comment.
Yes that's good now. The reason I didn't originally is just that it slows down the test a lot since we need to recompile for each slice.
jacobhinkle
commented
Sep 26, 2023
zasdfgbnm
reviewed
Sep 26, 2023
test/test_resize.cpp
Outdated
| fe.compileFusion(&fusion); | ||
|
|
||
| auto t0 = at::randn(shape, options); | ||
| for (auto [start, stop] : std::vector<std::pair<int64_t, int64_t>>( |
Collaborator
There was a problem hiding this comment.
Should we pull this set of slices out of the test and reuse it for all the three tests?
zasdfgbnm
approved these changes
Sep 26, 2023
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
This PR normalizes the inputs to
slicein order to mimic the semantics of numpy/PyTorch slicing. For an axis with extentext, if we receive a slice of(start, stop, step)we normalize it to(norm_start, norm_stop, step)whereSpecific changes in this PR:
sliceop.The simple Fusion in the input range test prints like this:
resulting in the following CUDA kernel:
This PR does NOT simplify these expressions for non-constant inputs. This can be done at concretization, which will be left for a follow-up PR.
Stacked on #892 and #895.
Fixes #439. Fixes #52.