Docs/AI Providers

AI Providers

MemryLab supports 9 LLM providers including 8 with free tiers. You can switch providers at any time without losing data. LLM and embedding providers can be configured independently — for example, use Gemini for analysis and Ollama for private local embeddings.

Recommended Local Models (Ollama)

For local inference, pick models based on your GPU VRAM. Embedding model is always nomic-embed-text.

VRAMLLM ModelSpeedInstall
4 GBllama3.2:3b~40 tok/sollama pull llama3.2:3b
8 GBllama3.1:8b~35 tok/sollama pull llama3.1:8b
12 GB ★qwen2.5:14b-instruct-q5_K_M~25 tok/sollama pull qwen2.5:14b-instruct-q5_K_M
16 GB+qwen2.5:32b-instruct-q4_K_M~15 tok/sollama pull qwen2.5:32b-instruct-q4_K_M

Larger models produce better belief extraction and contradiction detection. For no-GPU setups, use Gemini Flash (free) or Groq (free) as cloud providers.

💡 Pro Tip: Mix Providers

In Settings → Embedding Provider, you can use a different provider for embeddings vs LLM. The best privacy setup: use Gemini Flash (free, fast) for analysis + Ollama (local) for embeddings. This way your search index stays fully private while analysis uses a powerful cloud model.

Quick Comparison

ProviderFree TierDetails
Google GeminiYes1500 req/day on Gemini 1.5 Flash
GroqYesFree tier with rate limits
OpenRouterYesFree models available (Llama, Mistral, etc.)
CerebrasYesFree tier, fastest inference
MistralYesFree tier for small models
SambaNovaYesFree tier with generous limits
CohereYesFree trial API key
OpenAIPaidPay-per-use, starts at $0.002/1K tokens
Custom (OpenAI-compatible)PaidAny OpenAI-compatible endpoint (LM Studio, Ollama, etc.)

Google Gemini

1. Go to ai.google.dev and sign in
2. Create an API key
3. In MemryLab: Settings → LLM → Select "Gemini"
4. Paste your API key and test connection

https://ai.google.dev

Groq

1. Go to console.groq.com and create an account
2. Generate an API key from the dashboard
3. In MemryLab: Settings → LLM → Select "Groq"
4. Paste your API key and test connection

https://console.groq.com

OpenRouter

1. Go to openrouter.ai and create an account
2. Generate an API key
3. In MemryLab: Settings → LLM → Select "OpenRouter"
4. Paste your API key — free models are auto-selected

https://openrouter.ai

Cerebras

1. Go to cloud.cerebras.ai and sign up
2. Create an API key from the dashboard
3. In MemryLab: Settings → LLM → Select "Cerebras"
4. Paste your API key and test connection

https://cloud.cerebras.ai

Mistral

1. Go to console.mistral.ai and create an account
2. Generate an API key
3. In MemryLab: Settings → LLM → Select "Mistral"
4. Paste your API key and test connection

https://console.mistral.ai

SambaNova

1. Go to cloud.sambanova.ai and sign up
2. Create an API key
3. In MemryLab: Settings → LLM → Select "SambaNova"
4. Paste your API key and test connection

https://cloud.sambanova.ai

Cohere

1. Go to dashboard.cohere.com and sign up
2. Copy your trial API key
3. In MemryLab: Settings → LLM → Select "Cohere"
4. Paste your API key and test connection

https://dashboard.cohere.com

OpenAI

1. Go to platform.openai.com and sign in
2. Create an API key under API Keys
3. In MemryLab: Settings → LLM → Select "OpenAI"
4. Paste your API key and test connection

https://platform.openai.com

Custom (OpenAI-compatible)

1. In MemryLab: Settings → LLM → Select "Custom"
2. Enter the base URL (e.g., http://localhost:1234/v1)
3. Enter API key if required (or leave blank for local)
4. Select a model name and test connection

API Key Security

MemryLab stores API keys in your operating system's secure keychain — never in plain text files, environment variables, or the SQLite database.

  • Windows: Windows Credential Manager
  • macOS: Keychain Access
  • Linux: libsecret (GNOME Keyring / KWallet)