Skip to content

Tags: ServeurpersoCom/llama.cpp

Tags

b5335

Toggle b5335's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature.
CUDA: fix FlashAttention on Turing (ggml-org#13415)

b5334

Toggle b5334's commit message

Verified

This commit was created on GitHub.com and signed with GitHub’s verified signature.
arg : add env var to control mmproj (ggml-org#13416)

* arg : add env var to control mmproj

* small note about -hf --mmproj