forked from go-skynet/model-gallery
-
Notifications
You must be signed in to change notification settings - Fork 0
/
llama2-7b-chat-gguf.yaml
42 lines (37 loc) · 1.26 KB
/
llama2-7b-chat-gguf.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
name: "llama2-7b-chat-gguf"
description: |
Llama 2 7b chat in gguf format.
Llama 2 is a collection of pretrained and fine-tuned generative text models ranging in scale from 7 billion to 70 billion parameters..
license: "https://ai.meta.com/llama/license/"
urls:
- https://ai.meta.com/llama/
- https://huggingface.co/TheBloke/Llama-2-7b-Chat-GGUF
config_file: |
name: llama2-7b-chat-gguf
backend: "llama"
parameters:
top_k: 80
temperature: 0.2
top_p: 0.7
context_size: 4096
roles:
function: 'Function Result:'
assistant_function_call: 'Function Call:'
assistant: 'Assitant:'
user: 'User:'
system: 'System:'
template:
chat_message: llama2-7b-chat-gguf-chat
system_prompt: "You are a helpful assistant, below is a conversation, please respond with the next message and do not ask follow-up questions"
prompt_templates:
- name: "llama2-7b-chat-gguf-chat"
content: |
[INST]
{{if .SystemPrompt}}<<SYS>>{{.SystemPrompt}}<</SYS>>{{end}}
{{if .Input}}{{.Input}}{{end}}
[/INST]
Assistant:
files:
- filename: "llama-2-7b-chat.Q4_K_M.gguf"
sha256: "08a5566d61d7cb6b420c3e4387a39e0078e1f2fe5f055f3a03887385304d4bfa"
uri: "https://huggingface.co/TheBloke/Llama-2-7b-Chat-GGUF/resolve/main/llama-2-7b-chat.Q4_K_M.gguf"