Spaces:
Runtime error
Runtime error
Update pages/jarvis.py
Browse files- pages/jarvis.py +17 -20
pages/jarvis.py
CHANGED
@@ -77,32 +77,29 @@ def main():
|
|
77 |
chunk_size = st.slider("Chunk size", min_value=100, max_value=1000, value=600, step=20)
|
78 |
chunk_overlap = st.slider("Chunk overlap", min_value=10, max_value=200, value=40, step=10)
|
79 |
|
80 |
-
|
81 |
-
selected_tab = st.radio("Select Tab", tabs)
|
82 |
|
83 |
-
if
|
84 |
-
|
85 |
-
|
86 |
-
|
87 |
-
|
88 |
-
|
89 |
|
90 |
-
|
91 |
-
|
92 |
-
|
93 |
-
|
94 |
-
top_k = st.slider("Top-K Samples", min_value=1, max_value=10, value=3, step=1)
|
95 |
|
96 |
-
|
|
|
|
|
97 |
|
98 |
-
|
99 |
-
qa_chain = initialize_llmchain(llm_model, temperature, max_tokens, top_k, st.session_state["vector_db"])
|
100 |
-
|
101 |
-
elif selected_tab == "Chatbot":
|
102 |
-
if "qa_chain" in st.session_state:
|
103 |
message = st.text_input("Type your message")
|
104 |
if st.button("Submit"):
|
105 |
-
|
|
|
106 |
st.write("Assistant:", response["answer"])
|
107 |
|
108 |
if __name__ == "__main__":
|
|
|
77 |
chunk_size = st.slider("Chunk size", min_value=100, max_value=1000, value=600, step=20)
|
78 |
chunk_overlap = st.slider("Chunk overlap", min_value=10, max_value=200, value=40, step=10)
|
79 |
|
80 |
+
list_file_path = [file.name for file in uploaded_files]
|
|
|
81 |
|
82 |
+
if st.button("Generate Vector Database"):
|
83 |
+
st.text("Loading documents...")
|
84 |
+
doc_splits = load_doc(list_file_path, chunk_size, chunk_overlap)
|
85 |
+
st.text("Creating vector database...")
|
86 |
+
collection_name = create_collection_name(list_file_path[0])
|
87 |
+
vector_db = create_db(doc_splits, collection_name)
|
88 |
|
89 |
+
llm_model = st.selectbox("Choose LLM Model", list_llm)
|
90 |
+
temperature = st.slider("Temperature", min_value=0.01, max_value=1.0, value=0.7, step=0.1)
|
91 |
+
max_tokens = st.slider("Max Tokens", min_value=224, max_value=4096, value=1024, step=32)
|
92 |
+
top_k = st.slider("Top-K Samples", min_value=1, max_value=10, value=3, step=1)
|
|
|
93 |
|
94 |
+
if st.button("Initialize QA Chain"):
|
95 |
+
st.text("Initializing QA chain...")
|
96 |
+
qa_chain = initialize_llmchain(llm_model, temperature, max_tokens, top_k, vector_db)
|
97 |
|
98 |
+
st.header("Chatbot")
|
|
|
|
|
|
|
|
|
99 |
message = st.text_input("Type your message")
|
100 |
if st.button("Submit"):
|
101 |
+
st.text("Generating response...")
|
102 |
+
response = qa_chain({"question": message, "chat_history": []})
|
103 |
st.write("Assistant:", response["answer"])
|
104 |
|
105 |
if __name__ == "__main__":
|