These tools are available when you connect the LLMTest MCP server to your IDE. Your AI assistant calls them automatically based on your requests.
Show proxy status and activity summary. Returns flow count, total calls, spend, and pending suggestions.
List all AI flows with model, call count, latency, and cost per flow.
Get pending model-switch recommendations. Shows cost savings, latency differences, and quality comparisons.
Accept or dismiss a model suggestion.
id number
required
action string
required
Benchmark a flow against alternative models. The system selects the most relevant challengers based on your optimization goal, runs pairwise comparisons using an AI judge, and returns win/loss/tie records with cost and latency data.
flow string
required
currentModel string
optional
optimize_for "cost" | "quality" | "speed" | "balanced"
optional
challengers string[]
optional
Register test samples for a flow. Needed for pre-launch benchmarking when you don't have real traffic yet. The AI assistant generates realistic test prompts based on your description.
flow string
required
samples array
required
Show how many test samples are stored per flow and whether each flow is ready for benchmarking (needs at least 3).
Show new and trending models worth testing. Includes pricing, context length, and priority score.
Check your credit balance, total spend, and account info.
Submit feedback about the tool experience.
rating 1-5
required
comment string
optional