Skip to content
This repository was archived by the owner on Jun 4, 2025. It is now read-only.

Testmo#91

Merged
dhuangnm merged 4 commits intomainfrom
testmo
Aug 22, 2023
Merged

Testmo#91
dhuangnm merged 4 commits intomainfrom
testmo

Conversation

@dhuangnm
Copy link
Copy Markdown
Member

Improved GHA workflows to build nightly and release.

Added report build status to Testmo: https://neuralmagic.testmo.net/automation/sources/view/4

bfineran
bfineran previously approved these changes Aug 22, 2023
@dhuangnm dhuangnm merged commit 6167de3 into main Aug 22, 2023
bfineran pushed a commit that referenced this pull request Oct 26, 2023
* improve GHA workflow files to build nightly and release, and report status to testmo

* clean up

* report exit code

* Assign value to exit_code

---------

Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>
bfineran added a commit that referenced this pull request Oct 27, 2023
(previous commits)
* Add recipe_name to default file names

* Upgrade to transformers release V4.30.2 (#62)

* Update trainer and model flows to accommodate sparseml

Disable FP16 on QAT start (#12)

* Override LRScheduler when using LRModifiers

* Disable FP16 on QAT start

* keep wrapped scaler object for training after disabling

Using QATMatMul in DistilBERT model class (#41)

Removed double quantization of output of context layer. (#45)

Fix DataParallel validation forward signatures (#47)

* Fix: DataParallel validation forward signatures

* Update: generalize forward_fn selection

Best model after epoch (#46)

fix sclaer check for non fp16 mode in trainer (#38)

Mobilebert QAT (#55)

* Remove duplicate quantization of vocabulary.

enable a QATWrapper for non-parameterized matmuls in BERT self attention (#9)

* Utils and auxillary changes

update Zoo stub loading for SparseZoo 1.1 refactor (#54)

add flag to signal NM integration is active (#32)

Add recipe_name to file names

* Fix errors introduced in manual cherry-pick upgrade

Co-authored-by: Benjamin Fineran <bfineran@users.noreply.github.com>

* update build versions for NM fork pypi push (#74)

* fix nightly package name (#75)

* add make build command (#76)

* add GHA workflow files to build nightly and release packages (#77)

* add GHA workflow files to build nightly and release packages

* fix name

---------

Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>

* bump up version to 1.6.0 (#79)

Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>

---------

Co-authored-by: Konstantin <konstantin@neuralmagic.com>
Co-authored-by: Konstantin Gulin <66528950+KSGulin@users.noreply.github.com>
Co-authored-by: dhuangnm <74931910+dhuangnm@users.noreply.github.com>
Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>

minor improvements for build workflow files (#83)

Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>

fix minor issue (#84)

Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>

OPT with quantizable MatMuls (#85)

fix a minor issue for release build (#86)

Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>

update version in version.py

Testmo (#91)

* improve GHA workflow files to build nightly and release, and report status to testmo

* clean up

* report exit code

* Assign value to exit_code

---------

Co-authored-by: dhuang <dhuang@MacBook-Pro-2.local>

Update trainer.py - fix DistributedSampler import (#93)

DistributedSampler is used but not imported in `trainer.py`

Research/llama/bmm quantization (#94)

* Quantize attention matmuls

* Quantize attention matmuls

bump base transformers version
@dbogunowicz dbogunowicz deleted the testmo branch December 5, 2023 10:28
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants