Conversation
|
[For maintainers] Suggested jobs to run (before merge) run-slow: youtu |
|
run-slow: vit, clip, youtu |
|
This comment contains models: ["models/clip", "models/vit", "models/youtu"] |
|
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
|
View the CircleCI Test Summary for this PR: https://huggingface.co/spaces/transformers-community/circle-ci-viz?pr=45252&sha=49e459 |
|
I merged this PR to move fast, but if you think the output changes since your earlier PR #43734 for this model
|
* fix * fix --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
* fix * fix --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
|
Not sure about the batching, but for integration tests looks weird to me. Position ids should have remained the same after the PR, so I am going to check it (a bit later). We also had a few other big changes in generation/cache so there might have been a few PRs causing the diff 🤔 |
|
Ah the model was using |
* fix * fix --------- Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
What does this PR do?
#43166 used
torch.set_float32_matmul_precision("high")which causes (likely) TF32 being used(see this doc)
In #45248, we fixed an issue introduced #42428, which bring us back to the original setting: no TF32 used. This change works well and fixes all failing
test_batching_equivalence, if we apply the change in #45248 back to #42428.Somehow, even with the fix from #45248, the change
torch.set_float32_matmul_precision("high")in the filetests/models/youtu/test_modeling_youtu.pyin the later PR #43166 causes some failing tests (test_batching_equivalence, see this comment), because TF32 being used (or becausetreat each float32 number as the sum of two bfloat16 numbersis not good for us).This PR set it to
highest, i.e.which fixes all
test_batching_equivalenceagain, and achieve the goal mentioned by the PR author.The expected output values in the integration tests are updated, because the PR Prepare and keep track of position ids in
generate(#43734) changed the actual outputs. The new outputs look making sense and very similar to the previous ones, so I think it's safe.