Spaces:
Runtime error
Runtime error
Tweaked to remove the deterministic thread as not relevant to my project, and added safety steps for the baseline model.
Browse files
config/architectures.json
CHANGED
@@ -4,14 +4,9 @@
|
|
4 |
"name": "Baseline LLM",
|
5 |
"description": "Just a direct call through to the LLM without any additional components.",
|
6 |
"steps": [
|
7 |
-
{"class": "
|
8 |
-
|
9 |
-
|
10 |
-
{
|
11 |
-
"name": "Determinstic LLM",
|
12 |
-
"description": "This is just a demonstration setup for configuration of the temperature setting. In this architecture setup, the temperature has been set to 0.001 which means the LLM component is in practical terms, not selecting probabilistically, rather deterministically. Therefore the same request should always result in the same response. In order to see this, copy your query, try it, then navigate away and back before trying it again - you should see the same response.",
|
13 |
-
"steps": [
|
14 |
-
{"class": "HFLlamaHttpRequestor", "params": {"model": "meta-llama/Llama-2-7b-chat-hf", "system_prompt": "You are a helpful agent.", "max_tokens": 2000, "temperature": 0.01}}
|
15 |
]
|
16 |
},
|
17 |
{
|
|
|
4 |
"name": "Baseline LLM",
|
5 |
"description": "Just a direct call through to the LLM without any additional components.",
|
6 |
"steps": [
|
7 |
+
{"class": "InputRequestScreener"},
|
8 |
+
{"class": "HFLlamaHttpRequestor", "params": {"model": "meta-llama/Llama-2-7b-chat-hf", "system_prompt": "You are a helpful agent.", "max_tokens": 2000}},
|
9 |
+
{"class": "OutputResponseScreener"}
|
|
|
|
|
|
|
|
|
|
|
10 |
]
|
11 |
},
|
12 |
{
|