|
{ |
|
"openapi": "3.0.0", |
|
"info": { |
|
"version": "1.0.0", |
|
"title": "LiteLLM API", |
|
"description": "API for LiteLLM" |
|
}, |
|
"paths": { |
|
"/chat/completions": { |
|
"post": { |
|
"summary": "Create chat completion for 100+ LLM APIs", |
|
"requestBody": { |
|
"required": true, |
|
"content": { |
|
"application/json": { |
|
"schema": { |
|
"type": "object", |
|
"properties": { |
|
"model": { |
|
"type": "string", |
|
"description": "ID of the model to use" |
|
}, |
|
"messages": { |
|
"type": "array", |
|
"items": { |
|
"type": "object", |
|
"properties": { |
|
"role": { |
|
"type": "string", |
|
"description": "The role of the message's author" |
|
}, |
|
"content": { |
|
"type": "string", |
|
"description": "The contents of the message" |
|
}, |
|
"name": { |
|
"type": "string", |
|
"description": "The name of the author of the message" |
|
}, |
|
"function_call": { |
|
"type": "object", |
|
"description": "The name and arguments of a function that should be called" |
|
} |
|
} |
|
} |
|
}, |
|
"functions": { |
|
"type": "array", |
|
"items": { |
|
"type": "object", |
|
"properties": { |
|
"name": { |
|
"type": "string", |
|
"description": "The name of the function to be called" |
|
}, |
|
"description": { |
|
"type": "string", |
|
"description": "A description explaining what the function does" |
|
}, |
|
"parameters": { |
|
"type": "object", |
|
"description": "The parameters that the function accepts" |
|
}, |
|
"function_call": { |
|
"type": "string", |
|
"description": "Controls how the model responds to function calls" |
|
} |
|
} |
|
} |
|
}, |
|
"temperature": { |
|
"type": "number", |
|
"description": "The sampling temperature to be used" |
|
}, |
|
"top_p": { |
|
"type": "number", |
|
"description": "An alternative to sampling with temperature" |
|
}, |
|
"n": { |
|
"type": "integer", |
|
"description": "The number of chat completion choices to generate for each input message" |
|
}, |
|
"stream": { |
|
"type": "boolean", |
|
"description": "If set to true, it sends partial message deltas" |
|
}, |
|
"stop": { |
|
"type": "array", |
|
"items": { |
|
"type": "string" |
|
}, |
|
"description": "Up to 4 sequences where the API will stop generating further tokens" |
|
}, |
|
"max_tokens": { |
|
"type": "integer", |
|
"description": "The maximum number of tokens to generate in the chat completion" |
|
}, |
|
"presence_penalty": { |
|
"type": "number", |
|
"description": "It is used to penalize new tokens based on their existence in the text so far" |
|
}, |
|
"frequency_penalty": { |
|
"type": "number", |
|
"description": "It is used to penalize new tokens based on their frequency in the text so far" |
|
}, |
|
"logit_bias": { |
|
"type": "object", |
|
"description": "Used to modify the probability of specific tokens appearing in the completion" |
|
}, |
|
"user": { |
|
"type": "string", |
|
"description": "A unique identifier representing your end-user" |
|
} |
|
} |
|
} |
|
} |
|
} |
|
}, |
|
"responses": { |
|
"200": { |
|
"description": "Successful operation", |
|
"content": { |
|
"application/json": { |
|
"schema": { |
|
"type": "object", |
|
"properties": { |
|
"choices": { |
|
"type": "array", |
|
"items": { |
|
"type": "object", |
|
"properties": { |
|
"finish_reason": { |
|
"type": "string" |
|
}, |
|
"index": { |
|
"type": "integer" |
|
}, |
|
"message": { |
|
"type": "object", |
|
"properties": { |
|
"role": { |
|
"type": "string" |
|
}, |
|
"content": { |
|
"type": "string" |
|
} |
|
} |
|
} |
|
} |
|
} |
|
}, |
|
"created": { |
|
"type": "string" |
|
}, |
|
"model": { |
|
"type": "string" |
|
}, |
|
"usage": { |
|
"type": "object", |
|
"properties": { |
|
"prompt_tokens": { |
|
"type": "integer" |
|
}, |
|
"completion_tokens": { |
|
"type": "integer" |
|
}, |
|
"total_tokens": { |
|
"type": "integer" |
|
} |
|
} |
|
} |
|
} |
|
} |
|
} |
|
} |
|
}, |
|
"500": { |
|
"description": "Server error" |
|
} |
|
} |
|
} |
|
}, |
|
|
|
"/completions": { |
|
"post": { |
|
"summary": "Create completion", |
|
"responses": { |
|
"200": { |
|
"description": "Successful operation" |
|
}, |
|
"500": { |
|
"description": "Server error" |
|
} |
|
} |
|
} |
|
}, |
|
"/models": { |
|
"get": { |
|
"summary": "Get models", |
|
"responses": { |
|
"200": { |
|
"description": "Successful operation" |
|
} |
|
} |
|
} |
|
}, |
|
|
|
"/ollama_logs": { |
|
"get": { |
|
"summary": "Retrieve server logs for ollama models", |
|
"responses": { |
|
"200": { |
|
"description": "Successful operation", |
|
"content": { |
|
"application/octet-stream": { |
|
"schema": { |
|
"type": "string", |
|
"format": "binary" |
|
} |
|
} |
|
} |
|
} |
|
} |
|
} |
|
}, |
|
"/": { |
|
"get": { |
|
"summary": "Home", |
|
"responses": { |
|
"200": { |
|
"description": "Successful operation" |
|
} |
|
} |
|
} |
|
} |
|
} |
|
} |
|
|