doberst commited on
Commit
f567c65
·
verified ·
1 Parent(s): 785b9f2

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -2
README.md CHANGED
@@ -13,7 +13,12 @@ This model implements a generative 'question' and 'answer' (e.g., 'qa-gen') func
13
 
14
  `{'question': ['What was the amount of revenue in the quarter?'], 'answer': ['$3.2 billion']} `
15
 
16
- The model has been designed to accept one of three different parameters to guide the type of question-answer created: 'question, answer' (generates a standard question and answer), 'boolean' (generates a 'yes-no' question and answer), and 'multiple choice' (generates a multiple choice question and answer).
 
 
 
 
 
17
 
18
  [**slim-qa-gen-phi-3**](https://huggingface.co/llmware/slim-qa-gen-phi-3) is the Pytorch version of the model, and suitable for fine-tuning for further domain adaptation.
19
 
@@ -29,7 +34,7 @@ Load in your favorite GGUF inference engine, or try with llmware as follows:
29
  from llmware.models import ModelCatalog
30
 
31
  # to load the model and make a basic inference
32
- model = ModelCatalog().load_model("slim-qa-gen-tiny-tool", temperature=0.7, sample=True)
33
  response = model.function_call(text_sample, params=["boolean"])
34
 
35
  # this one line will download the model and run a series of tests
 
13
 
14
  `{'question': ['What was the amount of revenue in the quarter?'], 'answer': ['$3.2 billion']} `
15
 
16
+ The model has been designed to accept one of three different parameters to guide the type of question-answer created:
17
+ -- 'question, answer' (generates a standard question and answer),
18
+ -- 'boolean' (generates a 'yes-no' question and answer), and
19
+ -- 'multiple choice' (generates a multiple choice question and answer).
20
+
21
+ Note: we would generally recommend using sampling and temperature(0.5+) for varied generations, although if using 'multiple choice' mode, then we have seen the best results with temperature in the 0.2-0.3 range.
22
 
23
  [**slim-qa-gen-phi-3**](https://huggingface.co/llmware/slim-qa-gen-phi-3) is the Pytorch version of the model, and suitable for fine-tuning for further domain adaptation.
24
 
 
34
  from llmware.models import ModelCatalog
35
 
36
  # to load the model and make a basic inference
37
+ model = ModelCatalog().load_model("slim-qa-gen-phi-3-tool", temperature=0.5, sample=True)
38
  response = model.function_call(text_sample, params=["boolean"])
39
 
40
  # this one line will download the model and run a series of tests