Skip to main content

Why Gemini?

  • Free tier: 1,500 requests/day on gemini-2.0-flash — sufficient for most teams
  • 1M context window: Feed the full health history of all servers at once
  • No credit card required to get started

Setup

  1. Go to aistudio.google.com/app/apikey
  2. Click Create API key — free, no credit card needed
  3. Set the environment variable:
export GEMINI_API_KEY=AIza...
  1. Configure in .langsight.yaml:
investigate:
  provider: gemini
  model: gemini-2.0-flash    # default

Models

ModelContextFree tierBest for
gemini-2.0-flash1M1,500/dayDefault — fast, capable
gemini-2.5-pro1M50/dayBest quality analysis
gemini-1.5-flash1M1,500/dayBudget alternative

Free tier limits

ModelRequests/dayRequests/minute
gemini-2.0-flash1,50015
gemini-2.5-pro502
gemini-1.5-flash1,50015

Pricing (after free tier)

ModelInputOutput
gemini-2.0-flash$0.10/1M$0.40/1M
gemini-2.5-pro$1.25/1M$10/1M
gemini-1.5-flash$0.075/1M$0.30/1M

How it works

LangSight uses Gemini via its OpenAI-compatible API endpoint — no extra SDK needed. The openai package (already installed) is pointed at Google’s servers:
POST https://generativelanguage.googleapis.com/v1beta/openai/chat/completions