Skip to content

Add MXFP4 GGUF QuantizationType #1677

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 4 commits into from
Aug 7, 2025
Merged

Add MXFP4 GGUF QuantizationType #1677

merged 4 commits into from
Aug 7, 2025

Conversation

CISC
Copy link
Contributor

@CISC CISC commented Aug 7, 2025

Added in GPT-OSS PR ggml-org/llama.cpp#15091

@CISC
Copy link
Contributor Author

CISC commented Aug 7, 2025

Hmmm, I think this test is wrong, it check two different things against eachother:
https://github.com/huggingface/huggingface.js/actions/runs/16801643530/job/47613401870?pr=1677#step:7:95

@CISC
Copy link
Contributor Author

CISC commented Aug 7, 2025

It checks GGMLQuantizationType against GGUF_QUANT_ORDER which is GGMLFileQuantizationType, they are not equivalent:

export const GGUF_QUANT_ORDER: GGMLFileQuantizationType[] = [

@ngxson
Copy link
Member

ngxson commented Aug 7, 2025

The CI seems to be broken, but this is quite a trivial change so I guess it's ok to merge without CI being green

@ngxson ngxson merged commit e841a53 into huggingface:main Aug 7, 2025
3 of 4 checks passed
@CISC CISC deleted the gguf-mxfp4 branch August 7, 2025 20:49
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants