Models and AI Providers
GPT, Llama, and Qwen - oh my!
Last updated
GPT, Llama, and Qwen - oh my!
Last updated
Kiln can use essentially any LLM model from a number of providers:
Run locally with Ollama
Connect a cloud provider like OpenAI, Groq, OpenRouter, AWS, Fireworks and more. You provide your own API keys and we never have access to your dataset.
Connect to any OpenAI compatible server, like LiteLLM for vLLM
When you first run Kiln, the app will prompt you to setup one or more AI providers. You need at least one for the core features of Kiln to function.
We currently support the following AI providers: Ollama, OpenRouter, OpenAI, Groq, Fireworks.ai, AWS Bedrock, or any OpenAI compatible endpoing (like LiteLLM, Anthropic, HuggingFace and more).
If you want to add providers after initial setup, open Settings > AI Providers & Models
.
If you want to remove a provider, edit your ~/.kiln_ai/settings.yaml
and remove the provider's config.
Models come in several flavours, from very easy to use, to advanced methods for expert users:
Custom OpenAI compatible servers
LiteLLM - Anthropic, Huggingface, VertexAI, TogetherAI, and more.
These are models that have been tested to work with Kiln's various features. These are the easiest to use, and generally won't result in errors.
To use these models simply connect any AI provider from the Settings page. Once connected, you can select these model from the model dropdown on the Run screen. The dropdown will warn if you attempt to use a model that doesn't support a feature (like structured output).
You can request we add models here.
GPT 4o Mini
OpenAI, OpenRouter
✅︎
✅︎
✅︎
GPT 4o
OpenAI, OpenRouter
✅︎
✅︎
✅︎
Claude 3.5 Haiku
OpenRouter
✅︎
✅︎
Claude 3.5 Sonnet
OpenRouter
✅︎
✅︎
DeepSeek v3
OpenRouter
✅︎
✅︎
Gemini 1.5 Pro
OpenRouter
Gemini 1.5 Flash
OpenRouter
✅︎
Gemini 1.5 Flash 8B
OpenRouter
Nemotron 70B
OpenRouter
Llama 3.1 8B
Amazon Bedrock, Fireworks AI, Groq, Ollama, OpenRouter
✅︎ (some providers)
✅︎ (some providers)
✅︎
Llama 3.1 70B
Amazon Bedrock, Fireworks AI, Groq, Ollama, OpenRouter
✅︎ (some providers)
✅︎ (some providers)
✅︎
Llama 3.1 405B
Amazon Bedrock, Fireworks AI, Ollama, OpenRouter
✅︎
✅︎ (some providers)
Mistral Nemo
OpenRouter
✅︎
✅︎
Mistral Large
Amazon Bedrock, Ollama, OpenRouter
✅︎
✅︎
Llama 3.2 1B
Fireworks AI, Ollama, OpenRouter
✅︎
Llama 3.2 3B
Fireworks AI, Ollama, OpenRouter
✅︎
Llama 3.2 11B
Fireworks AI, Ollama, OpenRouter
✅︎ (some providers)
✅︎ (some providers)
Llama 3.2 90B
Fireworks AI, Ollama, OpenRouter
✅︎
✅︎
Llama 3.3 70B
Fireworks AI, Groq, Ollama, OpenRouter
✅︎ (some providers)
✅︎ (some providers)
Phi 3.5
Fireworks AI, Ollama, OpenRouter
Gemma 2 2B
Ollama
Gemma 2 9B
Ollama, OpenRouter
✅︎
Gemma 2 27B
Ollama, OpenRouter
✅︎
Mixtral 8x7B
Ollama, OpenRouter
✅︎ (some providers)
✅︎ (some providers)
Qwen 2.5 7B
Ollama, OpenRouter
✅︎ (some providers)
✅︎ (some providers)
Qwen 2.5 72B
Fireworks AI, Ollama, OpenRouter
✅︎ (some providers)
✅︎ (some providers)
Any Ollama model you have installed on your server will be available to use in Kiln. To add models, simply install them with the Ollama CLI ollama pull <model_name>
.
Some Ollama models are included/tested, and will automatically appear in the model dropdown. Any untested Ollama models will still appear in the dropdown, but in the "Untested" section.
If you want to use a model that is not in the list but is supported by one of our AI providers, you can use a custom model.
To use a custom model, click "Add Model" in the "AI Providers & Models" section of Settings.
These will appear in the "untested" section of the model dropdown.
If you have an OpenAI compatible server (LiteLLM, vLLM, etc.), you can use it in Kiln.
To do this, add a "Custom API" in the "AI Providers & Models" section of Settings.
All models supported by this API will appear in the "untested" section of the model dropdown.
Notes:
The API must support the /v1/models
endpoint, so Kiln can access the list of models.
Many Kiln tasks require structured (JSON) output. These can be hard to get working on custom servers, as each server/model pair usually needs some configuration to reliably produce structured output (tools vs json_mode vs json parsing, schema format, etc).
Kiln works with LiteLLM, an open source proxy which exposes an OpenAI compatible API to over 100 model providers. If your preferred provider isn't built in, try LiteLLM! Simply add your LiteLLM URL in the "Custom API" section of Settings.