move prompts to json file
This commit is contained in:
parent
0fc80e0f21
commit
93e0ffc9e0
2 changed files with 15 additions and 40 deletions
12
model_prompts.json
Normal file
12
model_prompts.json
Normal file
|
@ -0,0 +1,12 @@
|
||||||
|
{
|
||||||
|
"alpaca": "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{prompt}\n\n### Response:\n\n",
|
||||||
|
"oasst": "{prompt}",
|
||||||
|
"vicuna": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.\n\nUSER: {prompt}\nASSISTANT:",
|
||||||
|
"hermes": "### Instruction:\n{prompt}\n\n### Response:\n",
|
||||||
|
"gpt4": "### Instruction:\n{prompt}\n\n### Response:\n",
|
||||||
|
"qlora": "### Human: {prompt}\n### Assistant:",
|
||||||
|
"tulu": "\n{prompt}\n\n(include newline)",
|
||||||
|
"wizardlm-7b": "{prompt}\n\n### Response:",
|
||||||
|
"wizardlm-13b": "{prompt}\n\n### Response:",
|
||||||
|
"wizardlm-30b": "{prompt}\n\n### Response:"
|
||||||
|
}
|
43
template.py
43
template.py
|
@ -1,45 +1,8 @@
|
||||||
from difflib import SequenceMatcher
|
from difflib import SequenceMatcher
|
||||||
|
import json
|
||||||
|
|
||||||
model_prompts = {
|
with open("model_prompts.json", "r") as f:
|
||||||
"alpaca": """Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
model_prompts = json.load(f)
|
||||||
|
|
||||||
### Instruction:
|
|
||||||
{prompt}
|
|
||||||
|
|
||||||
### Response:
|
|
||||||
|
|
||||||
""",
|
|
||||||
"oasst": "<|prompter|>{prompt}<|endoftext|><|assistant|>",
|
|
||||||
"vicuna": """A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
|
|
||||||
|
|
||||||
USER: {prompt}
|
|
||||||
ASSISTANT:""",
|
|
||||||
"hermes": """### Instruction:
|
|
||||||
{prompt}
|
|
||||||
|
|
||||||
### Response:
|
|
||||||
""",
|
|
||||||
"gpt4": """### Instruction:
|
|
||||||
{prompt}
|
|
||||||
|
|
||||||
### Response:
|
|
||||||
""",
|
|
||||||
"qlora": """### Human: {prompt}
|
|
||||||
### Assistant:""",
|
|
||||||
"tulu": """<|user|>
|
|
||||||
{prompt}
|
|
||||||
<|assistant|>
|
|
||||||
(include newline)""",
|
|
||||||
"wizardlm-7b": """{prompt}
|
|
||||||
|
|
||||||
### Response:""",
|
|
||||||
"wizardlm-13b": """{prompt}
|
|
||||||
|
|
||||||
### Response:""",
|
|
||||||
"wizardlm-30b": """{prompt}
|
|
||||||
|
|
||||||
### Response:""",
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
def template(model, prompt):
|
def template(model, prompt):
|
||||||
|
|
Loading…
Reference in a new issue