Skip to content

Support Falcon H1 models #109

@younesbelkada

Description

@younesbelkada

Thank you for your great work - when trying to run the command:

docker model run hf.co/tiiuae/Falcon-H1-0.5B-Instruct-GGUF

I get:

Unable to find model 'hf.co/tiiuae/Falcon-H1-0.5B-Instruct-GGUF' locally. Pulling from the server.
Downloaded: 299.89 MB
Model pulled successfully
Interactive chat mode started. Type '/bye' to exit.
> Hi
Failed to generate a response: error response: status=500 body=unable to load runner: error waiting for runner to be ready: llama.cpp terminated unexpectedly: llama.cpp exit status: exit status 1
with output: llama_model_load: error loading model: error loading model architecture: unknown model architecture: 'falcon-h1'

This is because falcon-h1 has been only recently added into llama.cpp: ggml-org/llama.cpp#14534 - is there a way to update llama.cpp to its latest version in order to support recently added models?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions