doberst commited on
Commit
65c64f3
1 Parent(s): f1df212

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +98 -0
README.md CHANGED
@@ -1,3 +1,101 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+
5
+ # Model Card for Model ID
6
+
7
+ <!-- Provide a quick summary of what the model is/does. -->
8
+
9
+ dragon-llama-7b-0.1 part of the dRAGon ("Delivering RAG On Private Cloud") model series, RAG-instruct trained on top of a LLama-2 base model.
10
+
11
+ DRAGON models are fine-tuned with high-quality custom instruct datasets, designed for production quality use in RAG scenarios.
12
+
13
+
14
+ ### Benchmark Tests
15
+
16
+ Evaluated against the benchmark test: [RAG-Instruct-Benchmark-Tester](https://www.huggingface.co/datasets/llmware/rag_instruct_benchmark_tester)
17
+ Average of 2 Test Runs with 1 point for correct answer, 0.5 point for partial correct or blank / NF, 0.0 points for incorrect, and -1 points for hallucinations.
18
+
19
+ --**Accuracy Score**: **99.0** correct out of 100
20
+ --Not Found Classification: 95.0%
21
+ --Boolean: 82.5%
22
+ --Math/Logic: 70.0%
23
+ --Complex Questions (1-5): 4 (Low-Medium)
24
+ --Summarization Quality (1-5): 4 (Coherent, extractive)
25
+ --Hallucinations: No hallucinations observed in test runs.
26
+
27
+ For test run results (and good indicator of target use cases), please see the files ("core_rag_test" and "answer_sheet" in this repo).
28
+
29
+ ### Model Description
30
+
31
+ <!-- Provide a longer summary of what this model is. -->
32
+
33
+ - **Developed by:** llmware
34
+ - **Model type:** LLama-2
35
+ - **Language(s) (NLP):** English
36
+ - **License:** Apache 2.0
37
+ - **Finetuned from model:** Llama-2-7B-Base
38
+
39
+ ## Uses
40
+
41
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
+
43
+ The intended use of DRAGON models is two-fold:
44
+
45
+ 1. Provide high-quality RAG-Instruct models designed for fact-based, no "hallucination" question-answering in connection with an enterprise RAG workflow.
46
+
47
+ 2. DRAGON models are fine-tuned on top of leading base foundation models, generally in the 6-7B+ range, and purposefully rolled-out across multiple base models to provide choices and "drop-in" replacements for RAG specific use cases.
48
+
49
+ 3. DRAGON models were trained on the same principles as the BLING models, so generally, it should be easy to "upgrade" from a BLING model in testing to a DRAGON model in production.
50
+
51
+
52
+ ### Direct Use
53
+
54
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
55
+
56
+ DRAGON is designed for enterprise automation use cases, especially in knowledge-intensive industries, such as financial services,
57
+ legal and regulatory industries with complex information sources.
58
+
59
+ DRAGON models have been trained for common RAG scenarios, specifically: question-answering, key-value extraction, and basic summarization as the core instruction types
60
+ without the need for a lot of complex instruction verbiage - provide a text passage context, ask questions, and get clear fact-based responses.
61
+
62
+
63
+ ## Bias, Risks, and Limitations
64
+
65
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
66
+
67
+ Any model can provide inaccurate or incomplete information, and should be used in conjunction with appropriate safeguards and fact-checking mechanisms.
68
+
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ The fastest way to get started with BLING is through direct import in transformers:
73
+
74
+ from transformers import AutoTokenizer, AutoModelForCausalLM
75
+ tokenizer = AutoTokenizer.from_pretrained("dragon-llama-7b-0.1")
76
+ model = AutoModelForCausalLM.from_pretrained("dragon-llama-7b-0.1")
77
+
78
+
79
+ The BLING model was fine-tuned with a simple "\<human> and \<bot> wrapper", so to get the best results, wrap inference entries as:
80
+
81
+ full_prompt = "\<human>\: " + my_prompt + "\n" + "\<bot>\:"
82
+
83
+ The BLING model was fine-tuned with closed-context samples, which assume generally that the prompt consists of two sub-parts:
84
+
85
+ 1. Text Passage Context, and
86
+ 2. Specific question or instruction based on the text passage
87
+
88
+ To get the best results, package "my_prompt" as follows:
89
+
90
+ my_prompt = {{text_passage}} + "\n" + {{question/instruction}}
91
+
92
+
93
+
94
+ ## Model Card Contact
95
+
96
+ Darren Oberst & llmware team
97
+
98
+ Please reach out anytime if you are interested in this project!
99
+
100
+
101
+