{ "factory_llm_infos": [ { "name": "OpenAI", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION", "status": "1", "llm": [ { "llm_name": "gpt-4o-mini", "tags": "LLM,CHAT,128K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "gpt-4o", "tags": "LLM,CHAT,128K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "gpt-3.5-turbo", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "gpt-3.5-turbo-16k-0613", "tags": "LLM,CHAT,16k", "max_tokens": 16385, "model_type": "chat" }, { "llm_name": "text-embedding-ada-002", "tags": "TEXT EMBEDDING,8K", "max_tokens": 8191, "model_type": "embedding" }, { "llm_name": "text-embedding-3-small", "tags": "TEXT EMBEDDING,8K", "max_tokens": 8191, "model_type": "embedding" }, { "llm_name": "text-embedding-3-large", "tags": "TEXT EMBEDDING,8K", "max_tokens": 8191, "model_type": "embedding" }, { "llm_name": "whisper-1", "tags": "SPEECH2TEXT", "max_tokens": 26214400, "model_type": "speech2text" }, { "llm_name": "gpt-4", "tags": "LLM,CHAT,8K", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "gpt-4-turbo", "tags": "LLM,CHAT,8K", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "gpt-4-32k", "tags": "LLM,CHAT,32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "gpt-4-vision-preview", "tags": "LLM,CHAT,IMAGE2TEXT", "max_tokens": 765, "model_type": "image2text" } ] }, { "name": "Tongyi-Qianwen", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION", "status": "1", "llm": [ { "llm_name": "qwen-turbo", "tags": "LLM,CHAT,8K", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "qwen-plus", "tags": "LLM,CHAT,32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "text-embedding-v2", "tags": "TEXT EMBEDDING,2K", "max_tokens": 2048, "model_type": "embedding" }, { "llm_name": "paraformer-realtime-8k-v1", "tags": "SPEECH2TEXT", "max_tokens": 26214400, "model_type": "speech2text" }, { "llm_name": "qwen-vl-max", "tags": "LLM,CHAT,IMAGE2TEXT", "max_tokens": 765, "model_type": "image2text" }, { "llm_name": "qwen-vl-plus", "tags": "LLM,CHAT,IMAGE2TEXT", "max_tokens": 765, "model_type": "image2text" } ] }, { "name": "ZHIPU-AI", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION", "status": "1", "llm": [ { "llm_name": "glm-3-turbo", "tags": "LLM,CHAT,", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "glm-4", "tags": "LLM,CHAT,", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "glm-4v", "tags": "LLM,CHAT,IMAGE2TEXT", "max_tokens": 2000, "model_type": "image2text" }, { "llm_name": "embedding-2", "tags": "TEXT EMBEDDING", "max_tokens": 512, "model_type": "embedding" } ] }, { "name": "Ollama", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION", "status": "1", "llm": [] }, { "name": "LocalAI", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION", "status": "1", "llm": [] }, { "name": "OpenAI-API-Compatible", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION", "status": "1", "llm": [] }, { "name": "Moonshot", "logo": "", "tags": "LLM,TEXT EMBEDDING", "status": "1", "llm": [ { "llm_name": "moonshot-v1-8k", "tags": "LLM,CHAT,", "max_tokens": 7900, "model_type": "chat" }, { "llm_name": "moonshot-v1-32k", "tags": "LLM,CHAT,", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "moonshot-v1-128k", "tags": "LLM,CHAT", "max_tokens": 128000, "model_type": "chat" } ] }, { "name": "FastEmbed", "logo": "", "tags": "TEXT EMBEDDING", "status": "1", "llm": [ { "llm_name": "BAAI/bge-small-en-v1.5", "tags": "TEXT EMBEDDING,", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "BAAI/bge-small-zh-v1.5", "tags": "TEXT EMBEDDING,", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "BAAI/bge-base-en-v1.5", "tags": "TEXT EMBEDDING,", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "BAAI/bge-large-en-v1.5", "tags": "TEXT EMBEDDING,", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "sentence-transformers/all-MiniLM-L6-v2", "tags": "TEXT EMBEDDING,", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "nomic-ai/nomic-embed-text-v1.5", "tags": "TEXT EMBEDDING,", "max_tokens": 8192, "model_type": "embedding" }, { "llm_name": "jinaai/jina-embeddings-v2-small-en", "tags": "TEXT EMBEDDING,", "max_tokens": 2147483648, "model_type": "embedding" }, { "llm_name": "jinaai/jina-embeddings-v2-base-en", "tags": "TEXT EMBEDDING,", "max_tokens": 2147483648, "model_type": "embedding" } ] }, { "name": "Xinference", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION,TEXT RE-RANK", "status": "1", "llm": [] }, { "name": "Youdao", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION", "status": "1", "llm": [ { "llm_name": "maidalun1020/bce-embedding-base_v1", "tags": "TEXT EMBEDDING,", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "maidalun1020/bce-reranker-base_v1", "tags": "RE-RANK, 512", "max_tokens": 512, "model_type": "rerank" } ] }, { "name": "DeepSeek", "logo": "", "tags": "LLM", "status": "1", "llm": [ { "llm_name": "deepseek-chat", "tags": "LLM,CHAT,", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "deepseek-coder", "tags": "LLM,CHAT,", "max_tokens": 16385, "model_type": "chat" } ] }, { "name": "VolcEngine", "logo": "", "tags": "LLM, TEXT EMBEDDING", "status": "1", "llm": [ { "llm_name": "Skylark2-pro-32k", "tags": "LLM,CHAT,32k", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "Skylark2-pro-4k", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" } ] }, { "name": "BaiChuan", "logo": "", "tags": "LLM,TEXT EMBEDDING", "status": "1", "llm": [ { "llm_name": "Baichuan2-Turbo", "tags": "LLM,CHAT,32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "Baichuan2-Turbo-192k", "tags": "LLM,CHAT,192K", "max_tokens": 196608, "model_type": "chat" }, { "llm_name": "Baichuan3-Turbo", "tags": "LLM,CHAT,32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "Baichuan3-Turbo-128k", "tags": "LLM,CHAT,128K", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "Baichuan4", "tags": "LLM,CHAT,128K", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "Baichuan-Text-Embedding", "tags": "TEXT EMBEDDING", "max_tokens": 512, "model_type": "embedding" } ] }, { "name": "Jina", "logo": "", "tags": "TEXT EMBEDDING, TEXT RE-RANK", "status": "1", "llm": [ { "llm_name": "jina-reranker-v1-base-en", "tags": "RE-RANK,8k", "max_tokens": 8196, "model_type": "rerank" }, { "llm_name": "jina-reranker-v1-turbo-en", "tags": "RE-RANK,8k", "max_tokens": 8196, "model_type": "rerank" }, { "llm_name": "jina-reranker-v1-tiny-en", "tags": "RE-RANK,8k", "max_tokens": 8196, "model_type": "rerank" }, { "llm_name": "jina-colbert-v1-en", "tags": "RE-RANK,8k", "max_tokens": 8196, "model_type": "rerank" }, { "llm_name": "jina-embeddings-v2-base-en", "tags": "TEXT EMBEDDING", "max_tokens": 8196, "model_type": "embedding" }, { "llm_name": "jina-embeddings-v2-base-de", "tags": "TEXT EMBEDDING", "max_tokens": 8196, "model_type": "embedding" }, { "llm_name": "jina-embeddings-v2-base-es", "tags": "TEXT EMBEDDING", "max_tokens": 8196, "model_type": "embedding" }, { "llm_name": "jina-embeddings-v2-base-code", "tags": "TEXT EMBEDDING", "max_tokens": 8196, "model_type": "embedding" }, { "llm_name": "jina-embeddings-v2-base-zh", "tags": "TEXT EMBEDDING", "max_tokens": 8196, "model_type": "embedding" } ] }, { "name": "BAAI", "logo": "", "tags": "TEXT EMBEDDING, TEXT RE-RANK", "status": "1", "llm": [ { "llm_name": "BAAI/bge-large-zh-v1.5", "tags": "TEXT EMBEDDING,", "max_tokens": 1024, "model_type": "embedding" }, { "llm_name": "BAAI/bge-reranker-v2-m3", "tags": "RE-RANK,2k", "max_tokens": 2048, "model_type": "rerank" } ] }, { "name": "MiniMax", "logo": "", "tags": "LLM,TEXT EMBEDDING", "status": "1", "llm": [ { "llm_name": "abab6.5-chat", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "abab6.5s-chat", "tags": "LLM,CHAT,245k", "max_tokens": 245760, "model_type": "chat" }, { "llm_name": "abab6.5t-chat", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "abab6.5g-chat", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "abab5.5s-chat", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" } ] }, { "name": "Mistral", "logo": "", "tags": "LLM,TEXT EMBEDDING", "status": "1", "llm": [ { "llm_name": "open-mixtral-8x22b", "tags": "LLM,CHAT,64k", "max_tokens": 64000, "model_type": "chat" }, { "llm_name": "open-mixtral-8x7b", "tags": "LLM,CHAT,32k", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "open-mistral-7b", "tags": "LLM,CHAT,32k", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "mistral-large-latest", "tags": "LLM,CHAT,32k", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "mistral-small-latest", "tags": "LLM,CHAT,32k", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "mistral-medium-latest", "tags": "LLM,CHAT,32k", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "codestral-latest", "tags": "LLM,CHAT,32k", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "mistral-embed", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "embedding" } ] }, { "name": "Azure-OpenAI", "logo": "", "tags": "LLM,TEXT EMBEDDING,SPEECH2TEXT,MODERATION", "status": "1", "llm": [ { "llm_name": "azure-gpt-4o-mini", "tags": "LLM,CHAT,128K", "max_tokens": 128000, "model_type": "image2text" }, { "llm_name": "azure-gpt-4o", "tags": "LLM,CHAT,128K", "max_tokens": 128000, "model_type": "chat,image2text" }, { "llm_name": "azure-gpt-35-turbo", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "azure-gpt-35-turbo-16k", "tags": "LLM,CHAT,16k", "max_tokens": 16385, "model_type": "chat" }, { "llm_name": "azure-text-embedding-ada-002", "tags": "TEXT EMBEDDING,8K", "max_tokens": 8191, "model_type": "embedding" }, { "llm_name": "azure-text-embedding-3-small", "tags": "TEXT EMBEDDING,8K", "max_tokens": 8191, "model_type": "embedding" }, { "llm_name": "azure-text-embedding-3-large", "tags": "TEXT EMBEDDING,8K", "max_tokens": 8191, "model_type": "embedding" }, { "llm_name": "azure-whisper-1", "tags": "SPEECH2TEXT", "max_tokens": 26214400, "model_type": "speech2text" }, { "llm_name": "azure-gpt-4", "tags": "LLM,CHAT,8K", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "azure-gpt-4-turbo", "tags": "LLM,CHAT,8K", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "azure-gpt-4-32k", "tags": "LLM,CHAT,32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "azure-gpt-4-vision-preview", "tags": "LLM,CHAT,IMAGE2TEXT", "max_tokens": 765, "model_type": "image2text" } ] }, { "name": "Bedrock", "logo": "", "tags": "LLM,TEXT EMBEDDING", "status": "1", "llm": [ { "llm_name": "ai21.j2-ultra-v1", "tags": "LLM,CHAT,8k", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "ai21.j2-mid-v1", "tags": "LLM,CHAT,8k", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "cohere.command-text-v14", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "cohere.command-light-text-v14", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "cohere.command-r-v1:0", "tags": "LLM,CHAT,128k", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "cohere.command-r-plus-v1:0", "tags": "LLM,CHAT,128k", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "anthropic.claude-v2", "tags": "LLM,CHAT,100k", "max_tokens": 102400, "model_type": "chat" }, { "llm_name": "anthropic.claude-v2:1", "tags": "LLM,CHAT,200k", "max_tokens": 204800, "model_type": "chat" }, { "llm_name": "anthropic.claude-3-sonnet-20240229-v1:0", "tags": "LLM,CHAT,200k", "max_tokens": 204800, "model_type": "chat" }, { "llm_name": "anthropic.claude-3-5-sonnet-20240620-v1:0", "tags": "LLM,CHAT,200k", "max_tokens": 204800, "model_type": "chat" }, { "llm_name": "anthropic.claude-3-haiku-20240307-v1:0", "tags": "LLM,CHAT,200k", "max_tokens": 204800, "model_type": "chat" }, { "llm_name": "anthropic.claude-3-opus-20240229-v1:0", "tags": "LLM,CHAT,200k", "max_tokens": 204800, "model_type": "chat" }, { "llm_name": "anthropic.claude-instant-v1", "tags": "LLM,CHAT,100k", "max_tokens": 102400, "model_type": "chat" }, { "llm_name": "amazon.titan-text-express-v1", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "amazon.titan-text-premier-v1:0", "tags": "LLM,CHAT,32k", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "amazon.titan-text-lite-v1", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "meta.llama2-13b-chat-v1", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "meta.llama2-70b-chat-v1", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "meta.llama3-8b-instruct-v1:0", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta.llama3-70b-instruct-v1:0", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistral.mistral-7b-instruct-v0:2", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistral.mixtral-8x7b-instruct-v0:1", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "mistral.mistral-large-2402-v1:0", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistral.mistral-small-2402-v1:0", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "amazon.titan-embed-text-v2:0", "tags": "TEXT EMBEDDING", "max_tokens": 8192, "model_type": "embedding" }, { "llm_name": "cohere.embed-english-v3", "tags": "TEXT EMBEDDING", "max_tokens": 2048, "model_type": "embedding" }, { "llm_name": "cohere.embed-multilingual-v3", "tags": "TEXT EMBEDDING", "max_tokens": 2048, "model_type": "embedding" } ] }, { "name": "Gemini", "logo": "", "tags": "LLM,TEXT EMBEDDING,IMAGE2TEXT", "status": "1", "llm": [ { "llm_name": "gemini-1.5-pro-latest", "tags": "LLM,CHAT,1024K", "max_tokens": 1048576, "model_type": "chat" }, { "llm_name": "gemini-1.5-flash-latest", "tags": "LLM,CHAT,1024K", "max_tokens": 1048576, "model_type": "chat" }, { "llm_name": "gemini-1.0-pro", "tags": "LLM,CHAT,30K", "max_tokens": 30720, "model_type": "chat" }, { "llm_name": "gemini-1.0-pro-vision-latest", "tags": "LLM,IMAGE2TEXT,12K", "max_tokens": 12288, "model_type": "image2text" }, { "llm_name": "text-embedding-004", "tags": "TEXT EMBEDDING", "max_tokens": 2048, "model_type": "embedding" } ] }, { "name": "Groq", "logo": "", "tags": "LLM", "status": "1", "llm": [ { "llm_name": "gemma-7b-it", "tags": "LLM,CHAT,15k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "gemma2-9b-it", "tags": "LLM,CHAT,15k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "llama3-70b-8192", "tags": "LLM,CHAT,6k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "llama3-8b-8192", "tags": "LLM,CHAT,30k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mixtral-8x7b-32768", "tags": "LLM,CHAT,5k", "max_tokens": 32768, "model_type": "chat" } ] }, { "name": "OpenRouter", "logo": "", "tags": "LLM,IMAGE2TEXT", "status": "1", "llm": [ { "llm_name": "nousresearch/hermes-2-theta-llama-3-8b", "tags": "LLM CHAT 16K", "max_tokens": 16384, "model_type": "chat" }, { "llm_name": "alpindale/magnum-72b", "tags": "LLM CHAT 16K", "max_tokens": 16384, "model_type": "chat" }, { "llm_name": "google/gemma-2-9b-it", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "google/gemma-2-9b-it:free", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "sao10k/l3-stheno-8b", "tags": "LLM CHAT 32K", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "openrouter/flavor-of-the-week", "tags": "LLM CHAT 32K", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "ai21/jamba-instruct", "tags": "LLM CHAT 250K", "max_tokens": 256000, "model_type": "chat" }, { "llm_name": "nvidia/nemotron-4-340b-instruct", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "anthropic/claude-3.5-sonnet", "tags": "LLM IMAGE2TEXT 195K", "max_tokens": 200000, "model_type": "image2text" }, { "llm_name": "anthropic/claude-3.5-sonnet:beta", "tags": "LLM IMAGE2TEXT 195K", "max_tokens": 200000, "model_type": "image2text" }, { "llm_name": "sao10k/l3-euryale-70b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-medium-4k-instruct", "tags": "LLM CHAT 4K", "max_tokens": 4000, "model_type": "chat" }, { "llm_name": "cognitivecomputations/dolphin-mixtral-8x22b", "tags": "LLM CHAT 64K", "max_tokens": 65536, "model_type": "chat" }, { "llm_name": "qwen/qwen-2-72b-instruct", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "openchat/openchat-8b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistralai/mistral-7b-instruct", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "mistralai/mistral-7b-instruct-v0.3", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "nousresearch/hermes-2-pro-llama-3-8b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-mini-128k-instruct", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-mini-128k-instruct:free", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-medium-128k-instruct", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-medium-128k-instruct:free", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "neversleep/llama-3-lumimaid-70b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "google/gemini-flash-1.5", "tags": "LLM IMAGE2TEXT 2734K", "max_tokens": 2800000, "model_type": "image2text" }, { "llm_name": "perplexity/llama-3-sonar-small-32k-chat", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "perplexity/llama-3-sonar-small-32k-online", "tags": "LLM CHAT 28K", "max_tokens": 28000, "model_type": "chat" }, { "llm_name": "perplexity/llama-3-sonar-large-32k-chat", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "perplexity/llama-3-sonar-large-32k-online", "tags": "LLM CHAT 28K", "max_tokens": 28000, "model_type": "chat" }, { "llm_name": "deepseek/deepseek-chat", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "deepseek/deepseek-coder", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "openai/gpt-4o", "tags": "LLM IMAGE2TEXT 125K", "max_tokens": 128000, "model_type": "image2text" }, { "llm_name": "openai/gpt-4o-2024-05-13", "tags": "LLM IMAGE2TEXT 125K", "max_tokens": 128000, "model_type": "image2text" }, { "llm_name": "meta-llama/llama-3-8b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta-llama/llama-3-70b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta-llama/llama-guard-2-8b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "liuhaotian/llava-yi-34b", "tags": "LLM IMAGE2TEXT 4K", "max_tokens": 4096, "model_type": "image2text" }, { "llm_name": "allenai/olmo-7b-instruct", "tags": "LLM CHAT 2K", "max_tokens": 2048, "model_type": "chat" }, { "llm_name": "qwen/qwen-110b-chat", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "qwen/qwen-72b-chat", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "qwen/qwen-32b-chat", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "qwen/qwen-14b-chat", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "qwen/qwen-7b-chat", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "qwen/qwen-4b-chat", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "meta-llama/llama-3-8b-instruct:free", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "neversleep/llama-3-lumimaid-8b", "tags": "LLM CHAT 24K", "max_tokens": 24576, "model_type": "chat" }, { "llm_name": "neversleep/llama-3-lumimaid-8b:extended", "tags": "LLM CHAT 24K", "max_tokens": 24576, "model_type": "chat" }, { "llm_name": "snowflake/snowflake-arctic-instruct", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "fireworks/firellava-13b", "tags": "LLM IMAGE2TEXT 4K", "max_tokens": 4096, "model_type": "image2text" }, { "llm_name": "lynn/soliloquy-l3", "tags": "LLM CHAT 24K", "max_tokens": 24576, "model_type": "chat" }, { "llm_name": "sao10k/fimbulvetr-11b-v2", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta-llama/llama-3-8b-instruct:extended", "tags": "LLM CHAT 16K", "max_tokens": 16384, "model_type": "chat" }, { "llm_name": "meta-llama/llama-3-8b-instruct:nitro", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta-llama/llama-3-70b-instruct:nitro", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta-llama/llama-3-8b-instruct", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta-llama/llama-3-70b-instruct", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistralai/mixtral-8x22b-instruct", "tags": "LLM CHAT 64K", "max_tokens": 65536, "model_type": "chat" }, { "llm_name": "microsoft/wizardlm-2-8x22b", "tags": "LLM CHAT 64K", "max_tokens": 65536, "model_type": "chat" }, { "llm_name": "microsoft/wizardlm-2-7b", "tags": "LLM CHAT 32K", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "undi95/toppy-m-7b:nitro", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "mistralai/mixtral-8x22b", "tags": "LLM CHAT 64K", "max_tokens": 65536, "model_type": "chat" }, { "llm_name": "openai/gpt-4-turbo", "tags": "LLM IMAGE2TEXT 125K", "max_tokens": 128000, "model_type": "image2text" }, { "llm_name": "google/gemini-pro-1.5", "tags": "LLM IMAGE2TEXT 2734K", "max_tokens": 2800000, "model_type": "image2text" }, { "llm_name": "cohere/command-r-plus", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "databricks/dbrx-instruct", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "sophosympatheia/midnight-rose-70b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "cohere/command", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "cohere/command-r", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "anthropic/claude-3-haiku", "tags": "LLM IMAGE2TEXT 195K", "max_tokens": 200000, "model_type": "image2text" }, { "llm_name": "anthropic/claude-3-haiku:beta", "tags": "LLM IMAGE2TEXT 195K", "max_tokens": 200000, "model_type": "image2text" }, { "llm_name": "google/gemma-7b-it:nitro", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistralai/mixtral-8x7b-instruct:nitro", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "mistralai/mistral-7b-instruct:nitro", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "meta-llama/llama-2-70b-chat:nitro", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "gryphe/mythomax-l2-13b:nitro", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "anthropic/claude-3-opus", "tags": "LLM IMAGE2TEXT 195K", "max_tokens": 200000, "model_type": "image2text" }, { "llm_name": "anthropic/claude-3-sonnet", "tags": "LLM IMAGE2TEXT 195K", "max_tokens": 200000, "model_type": "image2text" }, { "llm_name": "anthropic/claude-3-opus:beta", "tags": "LLM IMAGE2TEXT 195K", "max_tokens": 200000, "model_type": "image2text" }, { "llm_name": "anthropic/claude-3-sonnet:beta", "tags": "LLM IMAGE2TEXT 195K", "max_tokens": 200000, "model_type": "image2text" }, { "llm_name": "mistralai/mistral-large", "tags": "LLM CHAT 32K", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "google/gemma-7b-it", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "google/gemma-7b-it:free", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "nousresearch/nous-hermes-2-mistral-7b-dpo", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta-llama/codellama-70b-instruct", "tags": "LLM CHAT 2K", "max_tokens": 2048, "model_type": "chat" }, { "llm_name": "recursal/eagle-7b", "tags": "LLM CHAT 9K", "max_tokens": 10000, "model_type": "chat" }, { "llm_name": "openai/gpt-3.5-turbo-0613", "tags": "LLM CHAT 4K", "max_tokens": 4095, "model_type": "chat" }, { "llm_name": "openai/gpt-4-turbo-preview", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "undi95/remm-slerp-l2-13b:extended", "tags": "LLM CHAT 6K", "max_tokens": 6144, "model_type": "chat" }, { "llm_name": "nousresearch/nous-hermes-2-mixtral-8x7b-dpo", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "nousresearch/nous-hermes-2-mixtral-8x7b-sft", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "mistralai/mistral-tiny", "tags": "LLM CHAT 32K", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "mistralai/mistral-small", "tags": "LLM CHAT 32K", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "mistralai/mistral-medium", "tags": "LLM CHAT 32K", "max_tokens": 32000, "model_type": "chat" }, { "llm_name": "austism/chronos-hermes-13b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "neversleep/noromaid-mixtral-8x7b-instruct", "tags": "LLM CHAT 8K", "max_tokens": 8000, "model_type": "chat" }, { "llm_name": "nousresearch/nous-hermes-yi-34b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "mistralai/mistral-7b-instruct-v0.2", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "cognitivecomputations/dolphin-mixtral-8x7b", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "google/gemini-pro", "tags": "LLM CHAT 89K", "max_tokens": 91728, "model_type": "chat" }, { "llm_name": "google/gemini-pro-vision", "tags": "LLM IMAGE2TEXT 44K", "max_tokens": 45875, "model_type": "image2text" }, { "llm_name": "mistralai/mixtral-8x7b", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "mistralai/mixtral-8x7b-instruct", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "rwkv/rwkv-5-world-3b", "tags": "LLM CHAT 9K", "max_tokens": 10000, "model_type": "chat" }, { "llm_name": "recursal/rwkv-5-3b-ai-town", "tags": "LLM CHAT 9K", "max_tokens": 10000, "model_type": "chat" }, { "llm_name": "togethercomputer/stripedhyena-nous-7b", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "togethercomputer/stripedhyena-hessian-7b", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "koboldai/psyfighter-13b-2", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "gryphe/mythomist-7b", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "openrouter/cinematika-7b", "tags": "LLM CHAT 8K", "max_tokens": 8000, "model_type": "chat" }, { "llm_name": "nousresearch/nous-capybara-7b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "nousresearch/nous-capybara-7b:free", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "openchat/openchat-7b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "openchat/openchat-7b:free", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "neversleep/noromaid-20b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "gryphe/mythomist-7b:free", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "intel/neural-chat-7b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "anthropic/claude-2", "tags": "LLM CHAT 195K", "max_tokens": 200000, "model_type": "chat" }, { "llm_name": "anthropic/claude-2.1", "tags": "LLM CHAT 195K", "max_tokens": 200000, "model_type": "chat" }, { "llm_name": "anthropic/claude-instant-1.1", "tags": "LLM CHAT 98K", "max_tokens": 100000, "model_type": "chat" }, { "llm_name": "anthropic/claude-2:beta", "tags": "LLM CHAT 195K", "max_tokens": 200000, "model_type": "chat" }, { "llm_name": "anthropic/claude-2.1:beta", "tags": "LLM CHAT 195K", "max_tokens": 200000, "model_type": "chat" }, { "llm_name": "teknium/openhermes-2.5-mistral-7b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "nousresearch/nous-capybara-34b", "tags": "LLM CHAT 195K", "max_tokens": 200000, "model_type": "chat" }, { "llm_name": "openai/gpt-4-vision-preview", "tags": "LLM IMAGE2TEXT 125K", "max_tokens": 128000, "model_type": "image2text" }, { "llm_name": "lizpreciatior/lzlv-70b-fp16-hf", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "undi95/toppy-m-7b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "alpindale/goliath-120b", "tags": "LLM CHAT 6K", "max_tokens": 6144, "model_type": "chat" }, { "llm_name": "undi95/toppy-m-7b:free", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "openrouter/auto", "tags": "LLM CHAT 195K", "max_tokens": 200000, "model_type": "chat" }, { "llm_name": "openai/gpt-3.5-turbo-1106", "tags": "LLM CHAT 16K", "max_tokens": 16385, "model_type": "chat" }, { "llm_name": "openai/gpt-4-1106-preview", "tags": "LLM CHAT 125K", "max_tokens": 128000, "model_type": "chat" }, { "llm_name": "huggingfaceh4/zephyr-7b-beta:free", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "google/palm-2-chat-bison-32k", "tags": "LLM CHAT 89K", "max_tokens": 91750, "model_type": "chat" }, { "llm_name": "google/palm-2-codechat-bison-32k", "tags": "LLM CHAT 89K", "max_tokens": 91750, "model_type": "chat" }, { "llm_name": "teknium/openhermes-2-mistral-7b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "open-orca/mistral-7b-openorca", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "gryphe/mythomax-l2-13b:extended", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "xwin-lm/xwin-lm-70b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "openai/gpt-3.5-turbo-instruct", "tags": "LLM CHAT 4K", "max_tokens": 4095, "model_type": "chat" }, { "llm_name": "mistralai/mistral-7b-instruct-v0.1", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "mistralai/mistral-7b-instruct:free", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "pygmalionai/mythalion-13b", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "openai/gpt-3.5-turbo-16k", "tags": "LLM CHAT 16K", "max_tokens": 16385, "model_type": "chat" }, { "llm_name": "openai/gpt-4-32k", "tags": "LLM CHAT 32K", "max_tokens": 32767, "model_type": "chat" }, { "llm_name": "openai/gpt-4-32k-0314", "tags": "LLM CHAT 32K", "max_tokens": 32767, "model_type": "chat" }, { "llm_name": "meta-llama/codellama-34b-instruct", "tags": "LLM CHAT 8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "phind/phind-codellama-34b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "nousresearch/nous-hermes-llama2-13b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "mancer/weaver", "tags": "LLM CHAT 8K", "max_tokens": 8000, "model_type": "chat" }, { "llm_name": "anthropic/claude-2.0", "tags": "LLM CHAT 98K", "max_tokens": 100000, "model_type": "chat" }, { "llm_name": "anthropic/claude-instant-1", "tags": "LLM CHAT 98K", "max_tokens": 100000, "model_type": "chat" }, { "llm_name": "anthropic/claude-1", "tags": "LLM CHAT 98K", "max_tokens": 100000, "model_type": "chat" }, { "llm_name": "anthropic/claude-1.2", "tags": "LLM CHAT 98K", "max_tokens": 100000, "model_type": "chat" }, { "llm_name": "anthropic/claude-instant-1.0", "tags": "LLM CHAT 98K", "max_tokens": 100000, "model_type": "chat" }, { "llm_name": "anthropic/claude-2.0:beta", "tags": "LLM CHAT 98K", "max_tokens": 100000, "model_type": "chat" }, { "llm_name": "anthropic/claude-instant-1:beta", "tags": "LLM CHAT 98K", "max_tokens": 100000, "model_type": "chat" }, { "llm_name": "undi95/remm-slerp-l2-13b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "google/palm-2-chat-bison", "tags": "LLM CHAT 25K", "max_tokens": 25804, "model_type": "chat" }, { "llm_name": "google/palm-2-codechat-bison", "tags": "LLM CHAT 19K", "max_tokens": 20070, "model_type": "chat" }, { "llm_name": "gryphe/mythomax-l2-13b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "meta-llama/llama-2-13b-chat", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "meta-llama/llama-2-70b-chat", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "openai/gpt-3.5-turbo", "tags": "LLM CHAT 16K", "max_tokens": 16385, "model_type": "chat" }, { "llm_name": "openai/gpt-3.5-turbo-0125", "tags": "LLM CHAT 16K", "max_tokens": 16385, "model_type": "chat" }, { "llm_name": "openai/gpt-3.5-turbo-0301", "tags": "LLM CHAT 4K", "max_tokens": 4095, "model_type": "chat" }, { "llm_name": "openai/gpt-4", "tags": "LLM CHAT 8K", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "openai/gpt-4-0314", "tags": "LLM CHAT 8K", "max_tokens": 8191, "model_type": "chat" }, { "llm_name": "01-ai/yi-large", "tags": "LLM CHAT 32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "01-ai/yi-34b-200k", "tags": "LLM CHAT 195K", "max_tokens": 200000, "model_type": "chat" }, { "llm_name": "01-ai/yi-34b-chat", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "01-ai/yi-34b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "01-ai/yi-6b", "tags": "LLM CHAT 4K", "max_tokens": 4096, "model_type": "chat" } ] }, { "name": "StepFun", "logo": "", "tags": "LLM", "status": "1", "llm": [ { "llm_name": "step-1-8k", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "step-1-32k", "tags": "LLM,CHAT,32k", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "step-1-128k", "tags": "LLM,CHAT,128k", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "step-1-256k", "tags": "LLM,CHAT,256k", "max_tokens": 262144, "model_type": "chat" }, { "llm_name": "step-1v-8k", "tags": "LLM,CHAT,IMAGE2TEXT", "max_tokens": 8192, "model_type": "image2text" } ] }, { "name": "NVIDIA", "logo": "", "tags": "LLM,TEXT EMBEDDING, TEXT RE-RANK", "status": "1", "llm": [ { "llm_name": "nvidia/nemotron-4-340b-reward", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "aisingapore/sea-lion-7b-instruct", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "databricks/dbrx-instruct", "tags": "LLM,CHAT,16K", "max_tokens": 16384, "model_type": "chat" }, { "llm_name": "google/gemma-7b", "tags": "LLM,CHAT,32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "google/gemma-2b", "tags": "LLM,CHAT,16K", "max_tokens": 16384, "model_type": "chat" }, { "llm_name": "google/gemma-2-9b-it", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "google/gemma-2-27b-it", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "google/recurrentgemma-2b", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "mediatek/breeze-7b-instruct", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta/llama2-70b", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "meta/llama3-8b", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "meta/llama3-70b", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-medium-128k-instruct", "tags": "LLM,CHAT,128K", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-medium-4k-instruct", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "microsoftphi-3-mini-128k-instruct", "tags": "LLM,CHAT,128K", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-mini-4k-instruct", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-small-128k-instruct", "tags": "LLM,CHAT,128K", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "microsoft/phi-3-small-8k-instruct", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistralai/mistral-7b-instruct", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "mistralai/mistral-7b-instruct-v0.3", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "mistralai/mixtral-8x7b-instruct", "tags": "LLM,CHAT,32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "mistralai/mixtral-8x22b-instruct", "tags": "LLM,CHAT,64K", "max_tokens": 65536, "model_type": "chat" }, { "llm_name": "mistralai/mistral-large", "tags": "LLM,CHAT,32K", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "nv-mistralai/mistral-nemo-12b-instruct", "tags": "LLM,CHAT,128K", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "nvidia/llama3-chatqa-1.5-70b", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "nvidia/llama3-chatqa-1.5-8b", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "nvidia/nemotron-4-340b-instruct", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "seallms/seallm-7b-v2.5", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "snowflake/arctic", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "upstage/solar-10.7b-instruct", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "baai/bge-m3", "tags": "TEXT EMBEDDING,8K", "max_tokens": 8192, "model_type": "embedding" }, { "llm_name": "nvidia/embed-qa-4", "tags": "TEXT EMBEDDING,512", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "nvidia/nv-embed-v1", "tags": "TEXT EMBEDDING,32K", "max_tokens": 32768, "model_type": "embedding" }, { "llm_name": "nvidia/nv-embedqa-e5-v5", "tags": "TEXT EMBEDDING,512", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "nvidia/nv-embedqa-mistral-7b-v2", "tags": "TEXT EMBEDDING,512", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "nvidia/nv-rerankqa-mistral-4b-v3", "tags": "RE-RANK,512", "max_tokens": 512, "model_type": "rerank" }, { "llm_name": "nvidia/rerank-qa-mistral-4b", "tags": "RE-RANK,512", "max_tokens": 512, "model_type": "rerank" }, { "llm_name": "snowflake/arctic-embed-l", "tags": "TEXT EMBEDDING,512", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "adept/fuyu-8b", "tags": "LLM,IMAGE2TEXT,4K", "max_tokens": 4096, "model_type": "image2text" }, { "llm_name": "google/deplot", "tags": "LLM,IMAGE2TEXT,4K", "max_tokens": 4096, "model_type": "image2text" }, { "llm_name": "google/paligemma", "tags": "LLM,IMAGE2TEXT,4K", "max_tokens": 4096, "model_type": "image2text" }, { "llm_name": "Iiuhaotian/Ilava-v1.6-34b", "tags": "LLM,IMAGE2TEXT,4K", "max_tokens": 4096, "model_type": "image2text" }, { "llm_name": "Iiuhaotian/Ilava-v1.6-mistral-7b", "tags": "LLM,IMAGE2TEXT,4K", "max_tokens": 4096, "model_type": "image2text" }, { "llm_name": "microsoft/kosmos-2", "tags": "LLM,IMAGE2TEXT,4K", "max_tokens": 4096, "model_type": "image2text" }, { "llm_name": "microsoft/phi-3-vision-128k-instruct", "tags": "LLM,IMAGE2TEXT,128K", "max_tokens": 131072, "model_type": "image2text" }, { "llm_name": "nvidia/neva-22b", "tags": "LLM,IMAGE2TEXT,4K", "max_tokens": 4096, "model_type": "image2text" } ] }, { "name": "LM-Studio", "logo": "", "tags": "LLM,TEXT EMBEDDING,IMAGE2TEXT", "status": "1", "llm": [] }, { "name": "cohere", "logo": "", "tags": "LLM,TEXT EMBEDDING, TEXT RE-RANK", "status": "1", "llm": [ { "llm_name": "command-r-plus", "tags": "LLM,CHAT,128k", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "command-r", "tags": "LLM,CHAT,128k", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "command", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "command-nightly", "tags": "LLM,CHAT,128k", "max_tokens": 131072, "model_type": "chat" }, { "llm_name": "command-light", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "command-light-nightly", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "embed-english-v3.0", "tags": "TEXT EMBEDDING", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "embed-english-light-v3.0", "tags": "TEXT EMBEDDING", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "embed-multilingual-v3.0", "tags": "TEXT EMBEDDING", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "embed-multilingual-light-v3.0", "tags": "TEXT EMBEDDING", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "embed-english-v2.0", "tags": "TEXT EMBEDDING", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "embed-english-light-v2.0", "tags": "TEXT EMBEDDING", "max_tokens": 512, "model_type": "embedding" }, { "llm_name": "embed-multilingual-v2.0", "tags": "TEXT EMBEDDING", "max_tokens": 256, "model_type": "embedding" }, { "llm_name": "rerank-english-v3.0", "tags": "RE-RANK,4k", "max_tokens": 4096, "model_type": "rerank" }, { "llm_name": "rerank-multilingual-v3.0", "tags": "RE-RANK,4k", "max_tokens": 4096, "model_type": "rerank" }, { "llm_name": "rerank-english-v2.0", "tags": "RE-RANK,512", "max_tokens": 8196, "model_type": "rerank" }, { "llm_name": "rerank-multilingual-v2.0", "tags": "RE-RANK,512", "max_tokens": 512, "model_type": "rerank" } ] }, { "name": "Lepton", "logo": "", "tags": "LLM", "status": "1", "llm": [ { "llm_name": "dolphin-mixtral-8x7b", "tags": "LLM,CHAT,32k", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "gemma-7b", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "llama3-1-8b", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "llama3-8b", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "llama2-13b", "tags": "LLM,CHAT,4K", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "llama3-1-70b", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "llama3-70b", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "llama3-1-405b", "tags": "LLM,CHAT,8k", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistral-7b", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "mistral-8x7b", "tags": "LLM,CHAT,8K", "max_tokens": 8192, "model_type": "chat" }, { "llm_name": "nous-hermes-llama2", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "openchat-3-5", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "toppy-m-7b", "tags": "LLM,CHAT,4k", "max_tokens": 4096, "model_type": "chat" }, { "llm_name": "wizardlm-2-7b", "tags": "LLM,CHAT,32k", "max_tokens": 32768, "model_type": "chat" }, { "llm_name": "wizardlm-2-8x22b", "tags": "LLM,CHAT,64K", "max_tokens": 65536, "model_type": "chat" } ] } ] }