Skip to content
Snippets Groups Projects
Commit e6832e69 authored by Max Kimmich's avatar Max Kimmich
Browse files

Use Llama3.1 as default for llamachat

parent f2fd6467
No related branches found
No related tags found
No related merge requests found
...@@ -240,14 +240,14 @@ class Llama(LLMModel): ...@@ -240,14 +240,14 @@ class Llama(LLMModel):
group.add_argument( group.add_argument(
"--llama-model", "--llama-model",
type=str, type=str,
default="QuantFactory/Meta-Llama-3-8B-Instruct-GGUF", default="QuantFactory/Meta-Llama-3.1-8B-Instruct-GGUF",
help="A pre-trained model from HF hub", help="A pre-trained model from HF hub",
), ),
group.add_argument( group.add_argument(
"--llama-model-file", "--llama-model-file",
type=str, type=str,
# TODO provide some help for selecting model files, and point user to set this argument if needed # TODO provide some help for selecting model files, and point user to set this argument if needed
default="Meta-Llama-3-8B-Instruct.Q5_K_M.gguf", default="Meta-Llama-3.1-8B-Instruct.Q8_0.gguf",
help="Specify the model file in case of a pre-trained model from HF hub, e.g., a specific quantized version", help="Specify the model file in case of a pre-trained model from HF hub, e.g., a specific quantized version",
), ),
group.add_argument( group.add_argument(
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment