训练文本生成
You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

97 lines
3.8 KiB

from typing import Optional
from dataclasses import dataclass
@dataclass
class Template:
name: str
def __post_init__(self):
assert hasattr(self, "_format_{}".format(self.name)), "Template {} does not exist.".format(self.name)
def get_prompt(self, query: str, history: Optional[list] = None, prefix: Optional[str] = "") -> str:
return getattr(self, "_format_{}".format(self.name))(query, history, prefix)
def _format_vanilla(self, query: str, history: Optional[list], prefix: Optional[str] = "") -> str:
r"""
Use for language model inference without histories.
"""
return query
def _format_alpaca(self, query: str, history: Optional[list], prefix: Optional[str] = "") -> str:
2 years ago
r"""
Supports: https://huggingface.co/tatsu-lab/alpaca-7b-wdiff
https://github.com/ymcui/Chinese-LLaMA-Alpaca
2 years ago
"""
if prefix:
prompt = prefix
else:
prompt = "Below is an instruction that describes a task. "
prompt += "Write a response that appropriately completes the request.\n\n"
if history:
for old_query, response in history:
prompt += "### Instruction:\n{}\n\n### Response:\n{}\n\n".format(old_query, response)
prompt += "### Instruction:\n{}\n\n### Response:\n".format(query)
return prompt
def _format_vicuna(self, query: str, history: Optional[list], prefix: Optional[str] = "") -> str:
2 years ago
r"""
Supports: https://huggingface.co/lmsys/vicuna-7b-delta-v1.1
https://huggingface.co/lmsys/vicuna-13b-delta-v1.1
"""
if prefix:
prompt = prefix
else:
prompt = "A chat between a curious user and an artificial intelligence assistant. "
prompt += "The assistant gives helpful, detailed, and polite answers to the user's questions. "
if history:
for old_query, response in history:
prompt += "USER: {} ASSISTANT: {}</s>".format(old_query, response)
prompt += "USER: {} ASSISTANT: ".format(query)
return prompt
def _format_belle(self, query: str, history: Optional[list], prefix: Optional[str] = "") -> str:
2 years ago
r"""
Supports: https://huggingface.co/BelleGroup/BELLE-LLaMA-EXT-13B
"""
prompt = prefix
if history:
for old_query, response in history:
prompt += "Human: {}\n\nBelle: {}\n\n".format(old_query, response)
prompt += "Human: {}\n\nBelle: ".format(query)
return prompt
2 years ago
def _format_linly(self, query: str, history: Optional[list], prefix: Optional[str] = "") -> str:
r"""
Supports: https://github.com/CVI-SZU/Linly
"""
prompt = prefix
if history:
for old_query, response in history:
prompt += "User: {}\nBot: {}\n".format(old_query, response)
prompt += "User: {}\nBot: ".format(query)
return prompt
def _format_billa(self, query: str, history: Optional[list], prefix: Optional[str] = "") -> str:
r"""
Supports: https://github.com/Neutralzz/BiLLa
"""
prompt = prefix
if history:
for old_query, response in history:
prompt += "Human: {}\nAssistant: {}\n".format(old_query, response)
prompt += "Human: {}\nAssistant: ".format(query)
return prompt
def _format_ziya(self, query: str, history: Optional[list], prefix: Optional[str] = "") -> str:
2 years ago
r"""
Supports: https://huggingface.co/IDEA-CCNL/Ziya-LLaMA-13B-v1
"""
prompt = prefix
if history:
for old_query, response in history:
prompt += "<human>:{}\n<bot>:{}\n".format(old_query, response)
prompt += "<human>:{}\n<bot>:".format(query)
return prompt