When to Use Custom LLMs
When to Use Custom LLMs
- For an open-source LLM, like Mixtral
- To update the context during the conversation
- To customize the messages before they’re sent to an LLM
Using an LLM provider
You’ll first want to POST your API key via the/credential
endpoint:
Using your server
To set up your server to act as the LLM, you’ll need to create an endpoint that is compatible with the OpenAI Client. For best results, your endpoint should also support streaming completions. If your server is making calls to an OpenAI compatble API, you can pipe the requests directly back in your response to Vapi. If you’d like your OpenAI-compatible endpoint to be authenticated, you can POST your server’s API key and URL via the/credential
endpoint:
custom-llm
model provider: