Spaces:
Running
Running
clean up
Browse files- .env.example +1 -0
- README.md +9 -6
- assets/crag-workflow.png +0 -0
- assets/rag-workflow.png +0 -0
- results/perf_data_nvidia_4080.xlsx +0 -0
- results/raw_data_nvidia_4080.xlsx +0 -0
.env.example
CHANGED
@@ -54,6 +54,7 @@ USING_TORCH_BFLOAT16=true
|
|
54 |
# HUGGINGFACE_MODEL_NAME_OR_PATH="meta-llama/Llama-2-13b-chat-hf"
|
55 |
# HUGGINGFACE_MODEL_NAME_OR_PATH="meta-llama/Llama-2-70b-chat-hf"
|
56 |
HUGGINGFACE_MODEL_NAME_OR_PATH="microsoft/Orca-2-7b"
|
|
|
57 |
|
58 |
STABLELM_MODEL_NAME_OR_PATH="OpenAssistant/stablelm-7b-sft-v7-epoch-3"
|
59 |
|
|
|
54 |
# HUGGINGFACE_MODEL_NAME_OR_PATH="meta-llama/Llama-2-13b-chat-hf"
|
55 |
# HUGGINGFACE_MODEL_NAME_OR_PATH="meta-llama/Llama-2-70b-chat-hf"
|
56 |
HUGGINGFACE_MODEL_NAME_OR_PATH="microsoft/Orca-2-7b"
|
57 |
+
# HUGGINGFACE_MODEL_NAME_OR_PATH="microsoft/Orca-2-13b"
|
58 |
|
59 |
STABLELM_MODEL_NAME_OR_PATH="OpenAssistant/stablelm-7b-sft-v7-epoch-3"
|
60 |
|
README.md
CHANGED
@@ -9,13 +9,14 @@ app_file: app.py
|
|
9 |
pinned: false
|
10 |
license: apache-2.0
|
11 |
---
|
12 |
-
# ChatPDF - Talk to Your PDF Files
|
13 |
|
14 |
-
|
|
|
|
|
15 |
|
16 |
## How it works
|
17 |
|
18 |
-
We're using an AI methodology, namely
|
19 |
|
20 |
At a very high level, the workflow can be divided into three stages:
|
21 |
|
@@ -25,7 +26,7 @@ At a very high level, the workflow can be divided into three stages:
|
|
25 |
|
26 |
3. Prompt execution / inference: Once the prompts have been compiled, they are submitted to a pre-trained LLM for inference—including both proprietary model APIs and open-source or self-trained models.
|
27 |
|
28 |
-

|
30 |
|
31 |
Tech stack used includes LangChain, Gradio, Chroma and FAISS.
|
32 |
- LangChain is an open-source framework that makes it easier to build scalable AI/LLM apps and chatbots.
|
|
|
102 |
# LLM_MODEL_TYPE=gpt4all-j
|
103 |
# LLM_MODEL_TYPE=gpt4all
|
104 |
# LLM_MODEL_TYPE=llamacpp
|
105 |
+
LLM_MODEL_TYPE=huggingface
|
106 |
# LLM_MODEL_TYPE=mosaicml
|
107 |
# LLM_MODEL_TYPE=stablelm
|
108 |
# LLM_MODEL_TYPE=openllm
|
109 |
+
# LLM_MODEL_TYPE=hftgi
|
110 |
```
|
111 |
|
112 |
- By default, the app runs `microsoft/orca-2-13b` model with HF Text Generation Interface, which runs on a research server and might be down from time to time.
|
|
|
124 |
# HUGGINGFACE_MODEL_NAME_OR_PATH="meta-llama/Llama-2-7b-chat-hf"
|
125 |
# HUGGINGFACE_MODEL_NAME_OR_PATH="meta-llama/Llama-2-13b-chat-hf"
|
126 |
# HUGGINGFACE_MODEL_NAME_OR_PATH="meta-llama/Llama-2-70b-chat-hf"
|
127 |
+
HUGGINGFACE_MODEL_NAME_OR_PATH="microsoft/Orca-2-7b"
|
128 |
+
# HUGGINGFACE_MODEL_NAME_OR_PATH="microsoft/Orca-2-13b"
|
129 |
```
|
130 |
|
assets/crag-workflow.png
DELETED
Binary file (255 kB)
|
|
assets/rag-workflow.png
ADDED
![]() |
results/perf_data_nvidia_4080.xlsx
DELETED
Binary file (5.87 kB)
|
|
results/raw_data_nvidia_4080.xlsx
DELETED
Binary file (25.8 kB)
|
|