Skip to content
Snippets Groups Projects
Commit d4bf2fd7 authored by Martin Forell's avatar Martin Forell
Browse files

ministral

parent e60da253
No related branches found
No related tags found
No related merge requests found
---
name: "dev-ministral"
config_file: |
context_size: 8192
f16: true
mmap: true
name: dev-ministral
parameters:
model: downloads/llama_backend/Ministral-8B-Instruct-2410-HF-f16.gguf
backend: llama
stopwords:
- "<|im_end|>"
- "<dummy32000>"
- "</tool_call>"
- "<|eot_id|>"
- "<|end_of_text|>"
- "</s>"
- "[/TOOL_CALLS]"
- "[/ACTIONS]"
- "[/INST]"
- "[INST]"
gpu_layers: 40
function:
# disable injecting the "answer" tool
disable_no_action: true
grammar:
# This allows the grammar to also return messages
#mixed_mode: true
# Not all models have a sketchpad or something to write thoughts on.
# This one will OR reply to strings OR JSON, but not both in the same reply
#no_mixed_free_string: true
# Disable grammar
# Base instructor model doesn't work well with grammars
disable: true
parallel_calls: true
disable_parallel_new_lines: true
return_name_in_function_response: true
# Without grammar uncomment the lines below
# Warning: this is relying only on the capability of the
# LLM model to generate the correct function call.
json_regex_match:
- "(?s)\\[TOOL\\_CALLS\\](.*)"
replace_function_results:
# Replace everything that is not JSON array or object
- key: '(?s)^[^{\[]*'
value: ""
- key: '(?s)[^}\]]*$'
value: ""
- key: "(?s)\\[TOOL\\_CALLS\\]"
value: ""
- key: "(?s)\\[\\/TOOL\\_CALLS\\]"
value: ""
template:
join_chat_messages_by_character: "" ## No newlines between messages
chat: |
{{.Input -}}
chat_message: |-
{{- if .FunctionCall -}}
[TOOL_CALLS] {{toJson .FunctionCall}} [/TOOL_CALLS]
{{- else if eq .RoleName "tool" -}}
[TOOL_RESULTS] {{.Content}} [/TOOL_RESULTS]
{{- else -}}
[INST] {{.Content }} [/INST]
{{ end -}}
completion: |
{{.Input}}
function: |-
[AVAILABLE_TOOLS] [{{range .Functions}}{"type": "function", "function": {"name": "{{.Name}}", "description": "{{.Description}}", "parameters": {{toJson .Parameters}} }}{{end}} ] [/AVAILABLE_TOOLS]{{.Input }}
files:
- filename: "downloads/llama_backend/Ministral-8B-Instruct-2410-HF-f16.gguf"
sha256: ""
uri: "https://huggingface.co/bartowski/Ministral-8B-Instruct-2410-HF-GGUF-TEST/resolve/main/Ministral-8B-Instruct-2410-HF-f16.gguf"
......@@ -169,6 +169,13 @@
- Text generation
- 8B Model
- functions
- url: https://gitlab.kit.edu/kit/aifb/BIS/infrastruktur/localai/localai-model-gallery/-/raw/main/dev-ministral.yaml
name: dev-ministral
tags:
- gpu
- Text generation
- 8B Model
- functions
- url: https://gitlab.kit.edu/kit/aifb/BIS/infrastruktur/localai/localai-model-gallery/-/raw/main/dev-codestral.yaml
name: dev-codestral
tags:
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment