Spaces:
Running
on
CPU Upgrade
Running
on
CPU Upgrade
from langchain.embeddings import OpenAIEmbeddings, HuggingFaceEmbeddings, HuggingFaceInferenceAPIEmbeddings | |
from langchain.vectorstores import Chroma, Qdrant | |
from qdrant_client import QdrantClient | |
from dotenv import load_dotenv | |
import os | |
provider_retrieval_model = "HF" | |
embeddingmodel = "BAAI/bge-small-en-v1.5" | |
load_dotenv() | |
HF_Token = os.environ.get("HF_TOKEN") | |
client_path = f"./vectorstore" | |
collection_name = f"collection" | |
provider_retrieval_model = "HF" | |
def create_vectorstore(docs): | |
if provider_retrieval_model == "HF": | |
qdrantClient = QdrantClient(path=client_path, prefer_grpc=True) | |
embeddings = HuggingFaceInferenceAPIEmbeddings( | |
api_key=HF_Token, model_name=embeddingmodel | |
) | |
dim = 1024 | |
qdrantClient.create_collection( | |
collection_name=collection_name, | |
vectors_config=VectorParams(size=dim, distance=Distance.COSINE), | |
) | |
vectorstore = Qdrant( | |
client=qdrantClient, | |
collection_name=collection_name, | |
embeddings=embeddings, | |
) | |
vectorstore.add_documents(docs_samp) |