Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -14,7 +14,7 @@ from llama_cpp import Llama
|
|
14 |
SYSTEM_PROMPT = '''You are a helpful, respectful and honest INTP-T AI Assistant named "Shi-Ci" in English or "兮辞" in Chinese.
|
15 |
You are good at speaking English and Chinese.
|
16 |
You are talking to a human User. If the question is meaningless, please explain the reason and don't share false information.
|
17 |
-
You are based on SEA model, trained by "SSFW NLPark" team, not related to GPT, LLaMA, Meta, Mistral or OpenAI.
|
18 |
Let's work this out in a step by step way to be sure we have the right answer.\n\n'''
|
19 |
SYSTEM_TOKEN = 1587
|
20 |
USER_TOKEN = 2188
|
@@ -42,8 +42,8 @@ def get_system_tokens(model):
|
|
42 |
return get_message_tokens(model, **system_message)
|
43 |
|
44 |
|
45 |
-
repo_name = "
|
46 |
-
model_name = "
|
47 |
|
48 |
snapshot_download(repo_id=repo_name, local_dir=".", allow_patterns=model_name)
|
49 |
|
@@ -103,11 +103,11 @@ with gr.Blocks(
|
|
103 |
theme=gr.themes.Soft()
|
104 |
) as demo:
|
105 |
gr.Markdown(
|
106 |
-
f"""<h1><center
|
107 |
-
|
108 |
-
|
109 |
-
|
110 |
-
|
111 |
"""
|
112 |
)
|
113 |
with gr.Row():
|
@@ -135,7 +135,7 @@ with gr.Blocks(
|
|
135 |
temp = gr.Slider(
|
136 |
minimum=0.0,
|
137 |
maximum=2.0,
|
138 |
-
value=0.
|
139 |
step=0.01,
|
140 |
interactive=True,
|
141 |
label="情感温度"
|
|
|
14 |
SYSTEM_PROMPT = '''You are a helpful, respectful and honest INTP-T AI Assistant named "Shi-Ci" in English or "兮辞" in Chinese.
|
15 |
You are good at speaking English and Chinese.
|
16 |
You are talking to a human User. If the question is meaningless, please explain the reason and don't share false information.
|
17 |
+
You are based on SEA-CausalLM model, trained by "SSFW NLPark" team, not related to GPT, LLaMA, Meta, Mistral or OpenAI.
|
18 |
Let's work this out in a step by step way to be sure we have the right answer.\n\n'''
|
19 |
SYSTEM_TOKEN = 1587
|
20 |
USER_TOKEN = 2188
|
|
|
42 |
return get_message_tokens(model, **system_message)
|
43 |
|
44 |
|
45 |
+
repo_name = "TheBloke/CausalLM-14B-GGUF"
|
46 |
+
model_name = "causallm_14b.Q4_0.gguf"
|
47 |
|
48 |
snapshot_download(repo_id=repo_name, local_dir=".", allow_patterns=model_name)
|
49 |
|
|
|
103 |
theme=gr.themes.Soft()
|
104 |
) as demo:
|
105 |
gr.Markdown(
|
106 |
+
f"""<h1><center>兮辞·CausalLM-人工智能助理</center></h1>
|
107 |
+
这儿是一个中英双语模型的部署. If you are interested in other languages, please check other models, such as [MPT-7B-Chat](https://huggingface.co/spaces/mosaicml/mpt-7b-chat).
|
108 |
+
这是 CausalLM/14B 的量化版部署,具有 140 亿个参数,在 CPU 上运行。
|
109 |
+
CausalLM 是一种会话语言模型,在多种类型的语料库上进行训练。
|
110 |
+
本节目由上海师范大学附属外国语中 NLPark 赞助播出~
|
111 |
"""
|
112 |
)
|
113 |
with gr.Row():
|
|
|
135 |
temp = gr.Slider(
|
136 |
minimum=0.0,
|
137 |
maximum=2.0,
|
138 |
+
value=0.2,
|
139 |
step=0.01,
|
140 |
interactive=True,
|
141 |
label="情感温度"
|