Skip to content
Snippets Groups Projects
Commit e60da253 authored by Martin Forell's avatar Martin Forell
Browse files

Update file dev-llama-3-large.yaml

parent 81df9d30
No related branches found
No related tags found
No related merge requests found
...@@ -9,7 +9,7 @@ config_file: | ...@@ -9,7 +9,7 @@ config_file: |
f16: true f16: true
gpu_layers: 85 gpu_layers: 85
parameters: parameters:
model: downloads/llama_backend/Llama-3.1-Nemotron-70B-Instruct-HF-Q6_K.gguf model: downloads/llama_backend/Llama-3.1-Nemotron-70B-Instruct-HF-Q5_K_S.gguf
template: template:
chat_message: | chat_message: |
...@@ -54,9 +54,9 @@ config_file: | ...@@ -54,9 +54,9 @@ config_file: |
"model": "dev-llama-3", "model": "dev-llama-3",
"messages": [{"role": "user", "content": "How are you doing?", "temperature": 0.1}] "messages": [{"role": "user", "content": "How are you doing?", "temperature": 0.1}]
}' }'
# files: files:
# - filename: "downloads/llama_backend/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf" - filename: "downloads/llama_backend/Llama-3.1-Nemotron-70B-Instruct-HF-Q5_K_S.gguf"
# sha256: "" sha256: ""
# uri: "https://huggingface.co/bartowski/Hermes-2-Theta-Llama-3-70B-GGUF/resolve/main/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf" uri: "https://huggingface.co/bartowski/Llama-3.1-Nemotron-70B-Instruct-HF-GGUF/resolve/main/Llama-3.1-Nemotron-70B-Instruct-HF-Q5_K_S.gguf"
# download model with huggingface-cli download bartowski/Llama-3.1-Nemotron-70B-Instruct-HF-GGUF --local-dir /data/localai-data/downloads/llama_backend --include '*Q6_K*' # download model with huggingface-cli download bartowski/Llama-3.1-Nemotron-70B-Instruct-HF-GGUF --local-dir /data/localai-data/downloads/llama_backend --include '*Q6_K*'
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment