Documentation Index
Fetch the complete documentation index at: https://docs.kaireonai.com/llms.txt
Use this file to discover all available pages before exploring further.
Overview
AI Configuration lets administrators tune the behavior of each AI analyzer and configure the LLM provider across the platform. Settings are saved per tenant (organization), so each organization can customize analysis thresholds, caps, algorithms, and AI provider independently. Navigate to Settings > AI Configuration in the KaireonAI sidebar.LLM Provider Configuration
The platform supports 6 LLM providers. Configuration is resolved in priority order:- Database — Per-tenant settings stored in
Settings > Integrations > AI / LLM Provider - Environment variables —
AI_PROVIDER,AI_MODEL,AI_API_KEY,AI_BASE_URL - Defaults — Google Gemini 2.5 Flash
| Provider | Key | Default Model | Base URL | Auth |
|---|---|---|---|---|
google | gemini-2.5-flash | Auto | API key | |
| Anthropic | anthropic | Claude models | Auto | API key |
| OpenAI | openai | GPT-4o models | Auto | API key |
| Amazon Bedrock | bedrock | Configurable | Auto | IAM role, access key, or inference profile |
| Ollama | ollama | Any local model | http://localhost:11434 | None |
| LM Studio | lm_studio | Any local model | http://localhost:1234/v1 | None (uses lm-studio placeholder key) |
access_key or iam_role), role ARN, access key ID, secret access key, and inference profile.
Analyzer Settings
Analyzer settings are managed via theGET /api/v1/ai/analyzer-settings and PUT /api/v1/ai/analyzer-settings endpoints. The GET endpoint is accessible to viewers, editors, and admins. The PUT endpoint requires the admin role.
Setup
Review the four analyzer sections
The page shows collapsible cards for each analyzer: Segmentation, Policy, Content Intelligence, and Rule Building. Each card lists the tunable parameters with their current values.
Adjust parameters
Change any value by editing the input field. Each parameter shows a plain-language description. Click Technical detail to see the underlying implementation detail.
Settings are loaded once on page load and edited locally until you save. Multiple admins editing at the same time will overwrite each other — the last save wins.
Segmentation Parameters
| Parameter | Type | Min | Max | Default | Description | Technical Detail |
|---|---|---|---|---|---|---|
| Min Clusters | integer | 2 | 20 | 2 | Fewest groups to split customers into | K-Means n_clusters lower bound. Fewer clusters = broader segments. |
| Max Clusters | integer | 2 | 50 | 8 | Most groups customers can be split into | K-Means n_clusters upper bound. More clusters = finer-grained segments. |
| Algorithm | enum | — | — | kmeans | Method used to find natural groupings | kmeans (centroid-based, fast) or dbscan (density-based, handles outliers). |
| Included Features | string[] or null | — | — | null (all) | Which attributes to consider when grouping | Feature whitelist for clustering input matrix. Null = all numeric features. |
Policy Parameters
| Parameter | Type | Min | Max | Default | Description | Technical Detail |
|---|---|---|---|---|---|---|
| Daily Cap | integer | 1 | 100 | 3 | Max messages a customer receives per day | Hard cap on contact frequency per customer per calendar day. |
| Weekly Cap | integer | 1 | 500 | 10 | Max messages a customer receives per week | Hard cap per customer per rolling 7-day window. |
| Monthly Cap | integer | 1 | 2,000 | 30 | Max messages a customer receives per month | Hard cap per customer per rolling 30-day window. |
| Lookback Days | integer | 7 | 730 | 90 | How far back to analyze interaction history | Interaction history window (days) for policy violation detection. |
| Min Sample Size | integer | 10 | 10,000 | 100 | Minimum interactions before analysis is meaningful | Statistical minimum sample size for reliable policy analysis. |
Content Intelligence Parameters
| Parameter | Type | Min | Max | Default | Description | Technical Detail |
|---|---|---|---|---|---|---|
| Min Impressions | integer | 1 | 100,000 | 100 | Times content must be shown before judging performance | Minimum impression count before metrics are statistically valid. |
| Metric Weights (CTR) | number | 0 | 1 | 0.33 | Weight given to click-through rate | Weighted blend coefficient for CTR in content scoring. |
| Metric Weights (CVR) | number | 0 | 1 | 0.34 | Weight given to conversion rate | Weighted blend coefficient for CVR in content scoring. |
| Metric Weights (Revenue) | number | 0 | 1 | 0.33 | Weight given to revenue per impression | Weighted blend coefficient for revenue-per-impression. |
| Confidence Level | number | 0.5 | 0.99 | 0.95 | How sure we need to be before recommending a change | Statistical confidence threshold (1 - alpha) for hypothesis tests. |
Rule Building Parameters
| Parameter | Type | Min | Max | Default | Description | Technical Detail |
|---|---|---|---|---|---|---|
| Max Conditions | integer | 1 | 20 | 5 | Maximum conditions allowed in a single rule | Upper bound on rule clause count to prevent overly complex rules. |
| Allowed Operators | string[] | — | — | equals, gt, lt, gte, lte, contains, in | Comparison types available when building rules | Operator whitelist for AI-generated rule conditions. |
| Field Type Constraints | string[] or null | — | — | null (all) | Which field types can be used in rules | Field type whitelist for rule condition targets. Null = no restrictions. |
Validation
All parameters are validated server-side using Zod schemas. If any value is outside its allowed range, the API returns a400 Bad Request with details about which parameter failed validation. Each sub-schema (segmentation, policy, content, ruleBuilder) is parsed independently — a bad value in one analyzer does not reset valid overrides in other analyzers.
Per-Run Overrides
Each AI analyzer (Segmentation, Policy Recommender, Content Intelligence, Rule Building) supports per-run parameter overrides in its Advanced Parameters panel. Per-run overrides:- Take effect for that run only and do not change the saved tenant configuration
- Default to the current tenant-level values from AI Configuration
- Include a Reset to Defaults button to restore tenant-level values
Large Dataset Warning Flow
When a dataset contains 5,000 or more rows (theML_ROW_THRESHOLD), the platform displays a confirmation dialog before proceeding with analysis. The dialog provides three pieces of information:
- Accuracy — The ML Worker uses K-Means clustering, logistic regression, and TF-IDF, which are more accurate than LLM pattern matching for large datasets.
- Cost estimate — Proceeding with the LLM shows the estimated token count and cost. The estimate is calculated as
rows x fields x 15 + 500tokens. - Speed — The ML Worker processes data locally in seconds vs. LLM round-trip latency.
- Use ML Worker — Routes the analysis to the ML Worker for full-dataset processing (only available if the ML Worker is connected and healthy).
- Proceed with LLM — Continues with LLM-based analysis using sampled data. Useful when the ML Worker is unavailable.
Next Steps
Auto-Segmentation
Configure segmentation analysis.
Policy Recommender
Optimize contact frequency policies.
Content Intelligence
Analyze creative performance.