From 588060fd86a1b3a35f21a1e24a5ebf49102619c2 Mon Sep 17 00:00:00 2001 From: Demian Frister <demian.frister@kit.edu> Date: Wed, 23 Oct 2024 07:44:20 +0000 Subject: [PATCH] =?UTF-8?q?hermes=20wieder=20hinzugef=C3=BCgt?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- dev-llama-3-large-hermes.yaml | 61 +++++++++++++++++++++++++++++++++++ index.yaml | 8 +++++ 2 files changed, 69 insertions(+) create mode 100644 dev-llama-3-large-hermes.yaml diff --git a/dev-llama-3-large-hermes.yaml b/dev-llama-3-large-hermes.yaml new file mode 100644 index 0000000..c929ade --- /dev/null +++ b/dev-llama-3-large-hermes.yaml @@ -0,0 +1,61 @@ +name: "dev-llama-3-large-hermes" + +description: | + This alias is for model with a high throughput. +config_file: | + name: dev-llama-3-large-hermes + mmap: true + backend: llama + f16: true + gpu_layers: 85 + parameters: + model: downloads/llama_backend/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf + + template: + chat_message: | + <|start_header_id|>{{if eq .RoleName "assistant"}}assistant{{else if eq .RoleName "system"}}system{{else if eq .RoleName "tool"}}tool{{else if eq .RoleName "user"}}user{{end}}<|end_header_id|> + + {{ if .FunctionCall -}} + Function call: + {{ else if eq .RoleName "tool" -}} + Function response: + {{ end -}} + {{ if .Content -}} + {{.Content -}} + {{ else if .FunctionCall -}} + {{ toJson .FunctionCall -}} + {{ end -}} + <|eot_id|> + function: | + <|start_header_id|>system<|end_header_id|> + + You are a function calling AI model. You are provided with function signatures within <tools></tools> XML tags. You may call one or more functions to assist with the user query. Don't make assumptions about what values to plug into functions. Here are the available tools: + <tools> + {{range .Functions}} + {'type': 'function', 'function': {'name': '{{.Name}}', 'description': '{{.Description}}', 'parameters': {{toJson .Parameters}} }} + {{end}} + </tools> + Use the following pydantic model json schema for each tool call you will make: + {'title': 'FunctionCall', 'type': 'object', 'properties': {'arguments': {'title': 'Arguments', 'type': 'object'}, 'name': {'title': 'Name', 'type': 'string'}}, 'required': ['arguments', 'name']}<|eot_id|><|start_header_id|>assistant<|end_header_id|> + Function call: + chat: | + <|begin_of_text|>{{.Input }} + <|start_header_id|>assistant<|end_header_id|> + completion: | + {{.Input}} + context_size: 8192 + stopwords: + - <|im_end|> + - <dummy32000> + - "<|eot_id|>" + - <|end_of_text|> + usage: | + curl http://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{ + "model": "dev-llama-3", + "messages": [{"role": "user", "content": "How are you doing?", "temperature": 0.1}] + }' +files: +- filename: "downloads/llama_backend/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf" + sha256: "" + uri: "https://huggingface.co/bartowski/Hermes-2-Theta-Llama-3-70B-GGUF/resolve/main/Hermes-2-Theta-Llama-3-70B-Q5_K_M.gguf" + \ No newline at end of file diff --git a/index.yaml b/index.yaml index 82cc249..06ba8fe 100644 --- a/index.yaml +++ b/index.yaml @@ -129,6 +129,14 @@ - Text generation - 70B Model - functions + +- url: https://gitlab.kit.edu/kit/aifb/BIS/infrastruktur/localai/localai-model-gallery/-/raw/main/dev-llama-3-large-hermes.yaml + name: dev-llama-3-large-hermes + tags: + - gpu + - Text generation + - 70B Model + - functions - url: https://gitlab.kit.edu/kit/aifb/BIS/infrastruktur/localai/localai-model-gallery/-/raw/main/dev-llama-3-small.yaml name: dev-llama-3-small -- GitLab