Skip to content

gguf-py: add support for I8, I16 and I32#6045

Merged
ggerganov merged 4 commits intoggml-org:masterfrom
certik:gguf_writer
Mar 14, 2024
Merged

gguf-py: add support for I8, I16 and I32#6045
ggerganov merged 4 commits intoggml-org:masterfrom
certik:gguf_writer

Conversation

@certik
Copy link
Copy Markdown
Contributor

@certik certik commented Mar 13, 2024

This code is equivalent as before, but now it is prepared to easily add
more NumPy dtypes.
@ggerganov
Copy link
Copy Markdown
Member

Will merge after #6050

@ggerganov ggerganov merged commit 3ca2348 into ggml-org:master Mar 14, 2024
ggerganov added a commit that referenced this pull request Mar 14, 2024
@certik certik deleted the gguf_writer branch March 14, 2024 14:13
@certik
Copy link
Copy Markdown
Contributor Author

certik commented Mar 14, 2024

@ggerganov my apologies for the tensor_shape vs tensor_dtype mistake --- I just discovered it this morning as well. I thought I tested it carefully, but I missed this. Thanks for fixing it!

@ggerganov
Copy link
Copy Markdown
Member

No problem. Btw, I think we need to update the gguf-py version

@certik
Copy link
Copy Markdown
Contributor Author

certik commented Mar 14, 2024

Btw, I think we need to update the gguf-py version

I sent a PR to do so here: #6060.

NeoZhangJianyu pushed a commit to NeoZhangJianyu/llama.cpp that referenced this pull request Mar 15, 2024
* Refactor dtype handling to be extensible

This code is equivalent as before, but now it is prepared to easily add
more NumPy dtypes.

* Add support for I8, I16 and I32

These types are allowed in the GGUF specification.

* Add support for I8, I16 and I32 to gguf_writer

* Add support for I8, I16, I32 to gguf_reader
NeoZhangJianyu pushed a commit to NeoZhangJianyu/llama.cpp that referenced this pull request Mar 15, 2024
mishig25 pushed a commit to huggingface/huggingface.js that referenced this pull request Jun 3, 2024
Bring `GGMLQuantizationType` up to date; adds `I8`, `I16`, `I32`, `I64`,
`F64`, `IQ1_M` and `BF16`.

Added in:
* ggml-org/llama.cpp#6045
* ggml-org/llama.cpp#6062
* ggml-org/llama.cpp#6302
* ggml-org/llama.cpp#6412
Seunghhon pushed a commit to Seunghhon/llama.cpp that referenced this pull request Apr 26, 2026
* Refactor dtype handling to be extensible

This code is equivalent as before, but now it is prepared to easily add
more NumPy dtypes.

* Add support for I8, I16 and I32

These types are allowed in the GGUF specification.

* Add support for I8, I16 and I32 to gguf_writer

* Add support for I8, I16, I32 to gguf_reader
Seunghhon pushed a commit to Seunghhon/llama.cpp that referenced this pull request Apr 26, 2026
phuongncn pushed a commit to phuongncn/llama.cpp-gx10-dgx-sparks-deepseekv4 that referenced this pull request Apr 28, 2026
* Refactor dtype handling to be extensible

This code is equivalent as before, but now it is prepared to easily add
more NumPy dtypes.

* Add support for I8, I16 and I32

These types are allowed in the GGUF specification.

* Add support for I8, I16 and I32 to gguf_writer

* Add support for I8, I16, I32 to gguf_reader
phuongncn pushed a commit to phuongncn/llama.cpp-gx10-dgx-sparks-deepseekv4 that referenced this pull request Apr 28, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants