Skip to content
Snippets Groups Projects
Commit 3db8e5e0 authored by Martin Forell's avatar Martin Forell
Browse files

new llama nvidia model

parent 39de119c
No related branches found
No related tags found
No related merge requests found
...@@ -9,7 +9,7 @@ config_file: | ...@@ -9,7 +9,7 @@ config_file: |
f16: true f16: true
gpu_layers: 85 gpu_layers: 85
parameters: parameters:
model: downloads/llama_backend/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf model: downloads/llama_backend/Llama-3.1-Nemotron-70B-Instruct-HF-Q6_K
template: template:
chat_message: | chat_message: |
...@@ -54,8 +54,9 @@ config_file: | ...@@ -54,8 +54,9 @@ config_file: |
"model": "dev-llama-3", "model": "dev-llama-3",
"messages": [{"role": "user", "content": "How are you doing?", "temperature": 0.1}] "messages": [{"role": "user", "content": "How are you doing?", "temperature": 0.1}]
}' }'
files: # files:
- filename: "downloads/llama_backend/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf" # - filename: "downloads/llama_backend/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf"
sha256: "" # sha256: ""
uri: "https://huggingface.co/bartowski/Hermes-2-Theta-Llama-3-70B-GGUF/resolve/main/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf" # uri: "https://huggingface.co/bartowski/Hermes-2-Theta-Llama-3-70B-GGUF/resolve/main/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf"
\ No newline at end of file # download model with huggingface-cli download bartowski/Llama-3.1-Nemotron-70B-Instruct-HF-GGUF --local-dir /data/localai-data/downloads/llama_backend --include '*Q6_K*'
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment