LLM Assist
LLM Assist is an AI-powered feature that reduces false positives by sending borderline findings to a large language model for contextual analysis. It acts as a second opinion on findings where pattern-based classifiers alone cannot determine with high confidence whether data is truly sensitive.
How It Works
- Finding Triage — After classifiers run, findings with confidence scores in the configurable ambiguous range are flagged for AI review.
- Context Extraction — The system extracts a minimal context window around each flagged finding (surrounding text, file metadata, column headers for structured data).
- AI Query — The context is sent to the AI model with a structured prompt asking whether the detected data is genuinely sensitive.
- Confidence Adjustment — Based on the AI’s response, the finding’s confidence score is adjusted up (confirmed) or down (dismissed).
- Result Merge — Adjusted findings are merged back into the scan results.
Multi-Provider Pipeline
Slim.io uses a multi-provider AI orchestrator with automatic failover. If the primary provider is unavailable or returns an error, the system automatically fails over to the next provider. This ensures LLM Assist remains available even during provider outages.
The multi-provider pipeline is fully managed by Slim.io. You do not need to provide API keys for any AI provider. All processing costs are included in the LLM Assist module pricing.
Deployment Modes
Slim-Hosted (Default)
In slim-hosted mode, data context is sent to Slim.io’s AI pipeline for analysis:
- Data leaves the customer’s cloud environment temporarily for AI processing
- Context windows are minimal (typically under 1 KB per finding)
- No data is stored in any provider’s systems (all providers configured with data processing agreements)
- Processing occurs in the same region as the Slim.io backend
Client-Hosted (BYOC)
In client-hosted mode, the AI model runs inside the customer’s own infrastructure:
- A containerized inference service is deployed in the customer’s VPC
- No data leaves the customer’s cloud boundary
- The customer provides the compute resources (GPU instances)
- Slim.io provides the model weights and inference container image
- Supported on AWS (SageMaker, ECS), GCP (Vertex AI, GKE), and Azure (AML, AKS)
Client-hosted LLM Assist requires GPU-capable instances and is significantly more expensive to operate than the slim-hosted option. It is recommended only for organizations with strict data residency requirements that prohibit any external data transfer.
Configuration
Confidence Thresholds
Configure which findings are sent to LLM Assist by adjusting the configurable ambiguous range. Findings below the floor are discarded, findings above the ceiling are accepted as-is, and findings inside the range are sent for AI review:
llm_assist:
enabled: true
min_confidence: <floor> # Findings below this are discarded without AI review
max_confidence: <ceiling> # Findings above this are accepted without AI review
max_findings_per_scan: 1000 # Cap on AI queries per scan jobThe default floor and ceiling are tuned by Slim.io and exposed via the dashboard UI under Settings > Detection. You can override them per environment.
Category Overrides
Override the configurable ambiguous range for specific PII categories — for example, more aggressive review for financial data, less aggressive for high-precision regex matches like email:
llm_assist:
category_overrides:
- category: Credit Card
min_confidence: <floor>
max_confidence: <ceiling>
- category: Email
min_confidence: <floor>
max_confidence: <ceiling>Accuracy Metrics
LLM Assist reports accuracy metrics per scan:
| Metric | Description |
|---|---|
| Findings Reviewed | Count of findings sent to AI |
| Confirmed | Findings where AI agreed with the classifier |
| Dismissed | Findings where AI determined the match was a false positive |
| Upgraded | Findings where AI increased the confidence score |
| Accuracy Rate | Percentage of AI decisions validated by human review (when available) |
These metrics are available in the Scan Monitor and via the API.
Cost Optimization
LLM Assist costs are driven by the number of findings sent for review. To optimize:
- Tune classifier confidence scores to reduce ambiguous findings
- Use proximity classifiers to increase base confidence
- Set
max_findings_per_scanto cap AI usage on large scans - Write suppression rules for known false positive patterns (they are filtered out before AI review)