client_llm: url: "" # <-- start your own inference endpoint and provide url here (or use https://api-inference.huggingface.co/models/HuggingFaceH4/zephyr-7b-beta) model_id: "HuggingFaceH4/zephyr-7b-beta" # <-- your client llm backend: HFChat max_tokens: 800 temperature: 0.6 expert_llm: url: "" # <-- start your own inference endpoint and provide url here (or use https://api-inference.huggingface.co/models/meta-llama/Meta-Llama-3-70B-Instruct) model_id: "meta-llama/Meta-Llama-3-70B-Instruct" backend: HFChat # <-- Currently supported: HFChat / VLLM / Fireworks classifier_llm: model_id: "MoritzLaurer/DeBERTa-v3-base-mnli-fever-anli" url: "" # <-- start your own inference endpoint of classifier model and provide url here batch_size: 8