Skip to content
This repository was archived by the owner on Jan 24, 2024. It is now read-only.

Conversation

@peakji
Copy link
Member

@peakji peakji commented Jun 9, 2023

No description provided.

@peakji peakji requested a review from fardeon June 9, 2023 02:58
@codecov-commenter
Copy link

Codecov Report

Patch coverage: 33.33% and project coverage change: -0.27 ⚠️

Comparison is base (0c4b8a5) 94.54% compared to head (2338e60) 94.27%.

❗ Your organization is not using the GitHub App Integration. As a result you may experience degraded service beginning May 15th. Please install the Github App Integration for your organization. Read more.

Additional details and impacted files
@@            Coverage Diff             @@
##           master     #209      +/-   ##
==========================================
- Coverage   94.54%   94.27%   -0.27%     
==========================================
  Files           7        7              
  Lines         330      332       +2     
==========================================
+ Hits          312      313       +1     
- Misses         18       19       +1     
Impacted Files Coverage Δ
basaran/model.py 88.46% <0.00%> (-0.58%) ⬇️
basaran/__init__.py 96.55% <100.00%> (+0.12%) ⬆️

☔ View full report in Codecov by Sentry.
📢 Do you have feedback about the report comment? Let us know in this issue.

@LoopControl
Copy link

LoopControl commented Jun 9, 2023

@peakji Thanks for working on this.

Just wanted to mention that while qlora 4 bit is a good option to have, it is around 8x slower in inference than GPTQ/AutoGPTQ.

I hope autogptq support is added one day in addition to this (also, gptq model download sizes are like a quarter of the size of full precision models so it can save a lot of space as well)

@peakji peakji merged commit 5e34a84 into master Jun 10, 2023
@peakji peakji deleted the load-in-4bit branch June 10, 2023 03:09
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants