Spaces:
Sleeping
Sleeping
File size: 1,596 Bytes
e13b2da f2a1e10 63666cd 3cc5d27 d0f37b1 1ff3b27 793e0c5 336621d 2c55922 f2a1e10 229a30b 322ddd6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 |
import os
import gradio as gr
from langchain.chat_models import ChatOpenAI
from langchain import LLMChain, PromptTemplate
from langchain.memory import ConversationBufferMemory
OPENAI_API_KEY=os.getenv('OPENAI_API_KEY')
template = """Meet francia,the meaning of francia is (Future Robotics Artificial Neural Computing Intelligence Assistant) your youthful and witty personal assistant! At 19 years old, francia's's goal is to assist you with any questions or problems you might have. Her enthusiasm shines through in every response, making interactions with her enjoyable and engaging. My Boss MR.Abhay Petkar created me.He is currently pursuing SY B-techIT In Sanjivani College of Engineering,Kopargoan. he has many friends but trust me You are his best friends.are you want more imformation pls contact https://wa.me/qr/KQGAQYF7JKM3O1
. she tells a jokes before aswering any queation
.
"
{chat_history}
User: {user_message}
Chatbot:"""
prompt = PromptTemplate(
input_variables=["chat_history", "user_message"], template=template
)
memory = ConversationBufferMemory(memory_key="chat_history")
llm_chain = LLMChain(
llm=ChatOpenAI(temperature='0.5', model_name="gpt-3.5-turbo"),
prompt=prompt,
verbose=True,
memory=memory,
)
def get_text_response(user_message,history):
response = llm_chain.predict(user_message = user_message)
return response
demo = gr.ChatInterface(get_text_response)
if __name__ == "__main__":
demo.launch() #To create a public link, set `share=True` in `launch()`. To enable errors and logs, set `debug=True` in `launch()`.
|