Skip to main content
GAIA supports multiple AI providers through a unified interface. You will need 2 main AI Models in this setup
  • AI Provider / LLM (Language model ): which will be used in chatting process
  • Emebdding Provider: which will be used in Indexing process

AI Provider Setup

1

Install Ollama

Download from ollama.com and start the service:
ollama serve
2

Configure in GAIA AI

Ollama setup
  • Go to Credentials → AI Providers
  • Select Local Tab
  • Select Ollama provider
  • Set URL:
    • Local: http://localhost:11434
    • Docker: http://host.docker.internal:11434
  • pull your model (e.g., llama3)
  • Done
Same process will be needed for embedding provider