Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -3,7 +3,8 @@ from langchain_core.prompts import ChatPromptTemplate, HumanMessagePromptTemplat
|
|
3 |
from langchain.schema import SystemMessage
|
4 |
import streamlit as st
|
5 |
import torch
|
6 |
-
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
|
|
7 |
from langchain_community.llms.huggingface_pipeline import HuggingFacePipeline
|
8 |
import nltk
|
9 |
import json
|
@@ -35,7 +36,7 @@ HF_TOKEN = os.getenv("HF_TOKEN")
|
|
35 |
|
36 |
# Model selection
|
37 |
model_name = st.selectbox('Choose a model:', [
|
38 |
-
'meta-llama/Llama-3.2-
|
39 |
'speakleash/Bielik-11B-v2.3-Instruct',
|
40 |
# Add your private model here
|
41 |
])
|
|
|
3 |
from langchain.schema import SystemMessage
|
4 |
import streamlit as st
|
5 |
import torch
|
6 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
7 |
+
from transformers.models.llama import LlamaTokenizer
|
8 |
from langchain_community.llms.huggingface_pipeline import HuggingFacePipeline
|
9 |
import nltk
|
10 |
import json
|
|
|
36 |
|
37 |
# Model selection
|
38 |
model_name = st.selectbox('Choose a model:', [
|
39 |
+
'meta-llama/Llama-3.2-3B-Instruct',
|
40 |
'speakleash/Bielik-11B-v2.3-Instruct',
|
41 |
# Add your private model here
|
42 |
])
|