Skip to content

scripts: add bpw per layer and model #14703

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 1 commit into from
Jul 15, 2025
Merged

Conversation

EAddario
Copy link
Contributor

Since llama-quantize allows users to select a wide range of quant types, sometimes it may not obvious to determine what's the most appropriate weight encoding scheme to use when following the GGUF naming conventions.

This PR modifies gguf_dump.py to display the bits per weight (bpw) for each layer, and for the overall model, when using the --markdown option.

@github-actions github-actions bot added the python python script changes label Jul 15, 2025
@CISC CISC merged commit c81f419 into ggml-org:master Jul 15, 2025
4 checks passed
@EAddario EAddario deleted the gguf_dump branch July 16, 2025 06:39
@EAddario
Copy link
Contributor Author

Thank you @CISC

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
python python script changes
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants