📄️ Pre-fix Assistant Messages
Supported by:
📄️ Predicted Outputs
| Property | Details |
📄️ Trimming Input Messages
Use litellm.trim_messages() to ensure messages does not exceed a model's token limit or specified max_tokens
📄️ Prompt Caching
Supported Providers:
📄️ Prompt Formatting
LiteLLM automatically translates the OpenAI ChatCompletions prompt format, to other models. You can control this by setting a custom prompt template for a model as well.