import gradio as gr from llama_cpp import Llama # Load model (gunakan versi GGUF yang ringan) model_path = "TheBloke/MythoMax-L2-13B-GGUF/mythomax.gguf" llm = Llama(model_path=model_path, n_ctx=2048) def chat(input_text): output = llm(input_text, max_tokens=100) return output["choices"][0]["text"] iface = gr.Interface( fn=chat, inputs="text", outputs="text", title="Dika AI - MythoMax Lite", description="Chatbot AI berbasis MythoMax 13B GGUF, optimized for Hugging Face CPU!" ) iface.launch()