move prompts to json file
This commit is contained in:
parent
0fc80e0f21
commit
93e0ffc9e0
2 changed files with 15 additions and 40 deletions
12
model_prompts.json
Normal file
12
model_prompts.json
Normal file
|
@ -0,0 +1,12 @@
|
|||
{
|
||||
"alpaca": "Below is an instruction that describes a task. Write a response that appropriately completes the request.\n\n### Instruction:\n{prompt}\n\n### Response:\n\n",
|
||||
"oasst": "{prompt}",
|
||||
"vicuna": "A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.\n\nUSER: {prompt}\nASSISTANT:",
|
||||
"hermes": "### Instruction:\n{prompt}\n\n### Response:\n",
|
||||
"gpt4": "### Instruction:\n{prompt}\n\n### Response:\n",
|
||||
"qlora": "### Human: {prompt}\n### Assistant:",
|
||||
"tulu": "\n{prompt}\n\n(include newline)",
|
||||
"wizardlm-7b": "{prompt}\n\n### Response:",
|
||||
"wizardlm-13b": "{prompt}\n\n### Response:",
|
||||
"wizardlm-30b": "{prompt}\n\n### Response:"
|
||||
}
|
43
template.py
43
template.py
|
@ -1,45 +1,8 @@
|
|||
from difflib import SequenceMatcher
|
||||
import json
|
||||
|
||||
model_prompts = {
|
||||
"alpaca": """Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
||||
|
||||
### Instruction:
|
||||
{prompt}
|
||||
|
||||
### Response:
|
||||
|
||||
""",
|
||||
"oasst": "<|prompter|>{prompt}<|endoftext|><|assistant|>",
|
||||
"vicuna": """A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.
|
||||
|
||||
USER: {prompt}
|
||||
ASSISTANT:""",
|
||||
"hermes": """### Instruction:
|
||||
{prompt}
|
||||
|
||||
### Response:
|
||||
""",
|
||||
"gpt4": """### Instruction:
|
||||
{prompt}
|
||||
|
||||
### Response:
|
||||
""",
|
||||
"qlora": """### Human: {prompt}
|
||||
### Assistant:""",
|
||||
"tulu": """<|user|>
|
||||
{prompt}
|
||||
<|assistant|>
|
||||
(include newline)""",
|
||||
"wizardlm-7b": """{prompt}
|
||||
|
||||
### Response:""",
|
||||
"wizardlm-13b": """{prompt}
|
||||
|
||||
### Response:""",
|
||||
"wizardlm-30b": """{prompt}
|
||||
|
||||
### Response:""",
|
||||
}
|
||||
with open("model_prompts.json", "r") as f:
|
||||
model_prompts = json.load(f)
|
||||
|
||||
|
||||
def template(model, prompt):
|
||||
|
|
Loading…
Reference in a new issue