Skip to content

top_p = 1 causes deterministic outputs #1797

@oobabooga

Description

@oobabooga

Setting top_p = 1 causes outputs to be identical even with a random seed. This was discovered by oobabooga/text-generation-webui#6431 (comment). See the full issue at oobabooga/text-generation-webui#6431.

Reproduction

from llama_cpp import Llama

# Load the model
model = Llama(
    model_path="models/Meta-Llama-3-8B-Instruct-Q4_K_S-HF/Meta-Llama-3-8B-Instruct-Q4_K_S.gguf",
    n_gpu_layers=128,
)

# Define the prompt
prompt = "Once upon a time"

for i in range(5):
    # Generate text with temperature = 1
    completion = model.create_completion(prompt=prompt, max_tokens=50, temperature=1.0, top_p=1.0, seed=-1)

    # Print the generated text
    print(completion['choices'][0]['text'])

The 5 outputs will be identical.

Verified with llama-cpp-python==0.3.1.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions