Skip to content

Add Agent Q Integration

Agent Q requires an LLM (Large Language Model) provider to power its AI capabilities. This guide walks you through the full setup from the moment you log in to Qualytics.

Prerequisites

Integration

Via Settings

Step 1: After logging in, click the Settings icon (gear) in the bottom-left sidebar.

home

Step 2: The Settings page opens on the Connections tab by default. Click the Integrations tab.

settings-connections

Step 3: Under the Agent Q section, click the Connect button next to LLM Configuration.

connect

Step 4: The LLM Configuration modal opens. Fill in the fields:

Field Description
Provider Select your LLM provider from the list (e.g., OpenAI, Anthropic, Google Gemini).
Model Choose a model available under the selected provider. You can also type a custom model name if your provider supports it.
API Key Enter the API key from your LLM provider. This is stored securely and never returned by the API.
Base URL (optional) Provide a custom endpoint URL for OpenAI-compatible providers (e.g., Ollama, OpenRouter, LiteLLM). Leave blank for standard providers.

modal-window

Step 5: Fill in your credentials and click Save to complete the configuration.

save

Info

When you save, Qualytics automatically validates your API key and tests the connection to the provider. If the key is invalid or the provider is unreachable, you will see an error before the configuration is stored. Qualytics also checks whether your provider supports web search — if it does, Agent Q can optionally search the Qualytics documentation to answer platform-related questions.

Once saved, Agent Q is ready to use. See How to Use Agent Q for next steps.

Via Agent Q Page

Step 1: Click Agent Q in the left sidebar to open the Agent Q page.

configured

Step 2: Click the Configure LLM button. You will be redirected to Settings > Integrations.

configured-llm

Step 3: Under the Agent Q section, click the Connect button next to LLM Configuration.

connect

Step 4: The LLM Configuration modal opens. Fill in the fields:

Field Description
Provider Select your LLM provider from the list (e.g., OpenAI, Anthropic, Google Gemini).
Model Choose a model available under the selected provider. You can also type a custom model name if your provider supports it.
API Key Enter the API key from your LLM provider. This is stored securely and never returned by the API.
Base URL (optional) Provide a custom endpoint URL for OpenAI-compatible providers (e.g., Ollama, OpenRouter, LiteLLM). Leave blank for standard providers.

modal-window

Step 5: Fill in your credentials and click Save to complete the configuration.

save

Once saved, Agent Q is ready to use. See How to Use Agent Q for next steps.

LLM Configuration per User

Each Qualytics user configures their own LLM provider independently. One user can use OpenAI while another uses Anthropic — both will have Agent Q available simultaneously using their respective configurations and API keys. Your API costs are your own responsibility.

Supported LLM Providers

Provider Example Models
OpenAI gpt-4o, gpt-4-turbo, o1, o3-mini
Anthropic claude-sonnet-4, claude-opus-4, claude-3-5-sonnet
Google Gemini gemini-2.0-flash, gemini-1.5-pro, gemini-1.5-flash
Amazon Bedrock Region-specific model IDs
Groq llama-3.3-70b, llama-3.1-8b, mixtral-8x7b
Mistral mistral-large, mistral-medium, codestral
Cohere command-r-plus, command-r
DeepSeek deepseek-chat, deepseek-coder
Ollama Any locally-hosted model (requires Base URL)
OpenRouter Any model via OpenRouter (requires Base URL)
LiteLLM Any model via LiteLLM proxy (requires Base URL)
Perplexity sonar-pro, sonar
Fireworks Any Fireworks-hosted model
GitHub Models GitHub-hosted models
Hugging Face Inference endpoint models
Together AI Any Together AI model
Cerebras llama-3.3-70b, llama-3.1-8b

Note

You must provide your own API key for the selected provider. Qualytics does not supply LLM API keys — you control your provider choice and associated costs.

What's Next?

Want to connect external AI clients like ChatGPT, Claude Desktop, Cursor, or VS Code directly to the Qualytics MCP server? See Connecting External AI Clients for step-by-step setup guides.