Skip to main content
HoopAI integrates leading large language models (LLMs) from providers such as OpenAI and Anthropic to power its suite of AI features. Rather than locking you into a single model, HoopAI automatically routes each task to the model best suited for the job — balancing speed, accuracy, cost, and capability behind the scenes so you can focus on running your business. This page explains which models power which features, how HoopAI evaluates and updates them, and what you can expect in terms of performance and limitations.

Models behind HoopAI features

HoopAI does not expose raw model names in the interface. Instead, the platform abstracts model selection into feature-level capabilities. Under the hood, every AI-powered feature is mapped to a specific model (or combination of models) that has been tested and validated for that use case.

Why HoopAI manages model selection

  • Reliability — Each feature is matched to a model that consistently meets quality thresholds for that task.
  • Cost efficiency — Simpler tasks use lighter models to keep your usage costs low, while complex tasks use more capable models.
  • Seamless upgrades — When a better model becomes available, HoopAI can switch transparently without breaking your workflows.

Capabilities matrix

The table below maps each major AI feature to the model tier that powers it, along with key characteristics.
FeatureModel tierPrimary providerInput typesTypical response timeContext window
Conversation AIAdvanced LLMOpenAI / AnthropicText, images1 — 4 secondsUp to 128K tokens
Voice AISpeech-optimized LLM + TTS/STTOpenAI / AnthropicAudio (real-time)Sub-second streamingUp to 32K tokens
Content AIAdvanced LLMOpenAIText, images2 — 8 secondsUp to 128K tokens
Reviews AIMid-tier LLMOpenAIText1 — 3 secondsUp to 16K tokens
Workflow AIAdvanced LLMOpenAI / AnthropicText, structured data1 — 5 secondsUp to 128K tokens
Model tiers and providers may change as HoopAI continuously evaluates the best options. The platform always selects models that meet or exceed the performance benchmarks for each feature.

Model tier definitions

  • Advanced LLM — The most capable general-purpose language models, used for tasks that require nuanced understanding, long context, or complex reasoning.
  • Mid-tier LLM — Highly capable models optimized for speed and cost when the task is well-scoped (for example, generating a review response from a template).
  • Speech-optimized LLM + TTS/STT — A combination of a language model for dialogue management, a speech-to-text (STT) engine for transcription, and a text-to-speech (TTS) engine for voice synthesis.

How HoopAI selects and evaluates models

HoopAI’s AI engineering team follows a structured evaluation process whenever a new model is released or an existing model is updated.
1

Benchmark testing

Every candidate model is tested against a curated set of benchmarks specific to each feature — appointment booking accuracy for Voice AI, factual grounding for Conversation AI, tone consistency for Content AI, and so on.
2

Quality scoring

Results are scored across five dimensions: accuracy, latency, cost per request, safety, and consistency. A model must meet minimum thresholds in all five categories to be considered.
3

Shadow deployment

Qualifying models run in shadow mode alongside the current production model. Real (anonymized) traffic is processed by both, and outputs are compared without affecting end users.
4

Gradual rollout

Once a model demonstrates equal or better performance, it is rolled out incrementally — starting with a small percentage of traffic and scaling to full deployment over several days.
5

Monitoring and rollback

Post-deployment monitoring tracks error rates, latency percentiles, and user satisfaction signals. If any metric degrades, HoopAI can roll back to the previous model within minutes.

Model update schedule and versioning

HoopAI evaluates new model versions on an ongoing basis. Major updates typically follow this cadence:
Update typeFrequencyUser impact
Minor model patchesMonthly or as released by providersNo user action required; improvements are automatic
Major model upgradesQuarterly evaluation cycleMay include new capabilities; announced in release notes
Provider additionsAs new providers meet quality barExpands the pool of models HoopAI can route to
You do not need to take any action when models are updated. HoopAI handles versioning and migration automatically. If a model change affects prompt behavior, the platform will notify you through the dashboard.

Performance characteristics

Response time ranges

Response times depend on the feature, the length of input, and the complexity of the task. The ranges below represent the 50th to 95th percentile under normal operating conditions.
Featurep50 latencyp95 latency
Conversation AI (text)1.2 seconds3.5 seconds
Voice AI (turn-by-turn)0.4 seconds1.2 seconds
Content AI (short-form)1.5 seconds4.0 seconds
Content AI (long-form)4.0 seconds12.0 seconds
Reviews AI0.8 seconds2.5 seconds
Workflow AI (GPT action)1.0 seconds4.5 seconds

Accuracy and quality

Accuracy metrics vary by feature:
  • Conversation AI — Measured by intent recognition accuracy (target above 92%) and factual grounding against knowledge base sources.
  • Voice AI — Measured by speech recognition word-error rate (target below 8%) and task-completion rate (target above 85%).
  • Content AI — Measured by brand-voice consistency scores and human quality ratings during internal evaluations.
  • Reviews AI — Measured by sentiment-match accuracy and response appropriateness ratings.
  • Workflow AI — Measured by structured output validity (JSON schema compliance) and action-execution success rates.

Limitations

Every AI model has inherent limitations. Understanding these helps you set appropriate expectations and design better prompts. For detailed prompt-writing strategies, see Prompt engineering 101.

Text model limitations

  • Hallucination — Models can generate plausible-sounding but incorrect information. Always provide a knowledge base so the model has grounded facts to reference.
  • Context window limits — Very long conversations may exceed the model’s context window, causing earlier messages to be summarized or dropped. Keep prompts concise and use guardrails to prevent unbounded conversations.
  • Recency — Models have a training data cutoff date and may not know about very recent events unless information is provided through your knowledge base or custom actions.
  • Sensitive topics — Models include built-in safety filters. In rare cases, these filters may decline to respond to legitimate business queries. Adjusting your prompt wording usually resolves this.

Voice model limitations

  • Background noise — Speech recognition accuracy decreases in noisy environments. Callers on speakerphone or in public spaces may experience lower transcription quality.
  • Accents and dialects — While modern STT engines handle a wide range of accents, recognition accuracy can vary. Testing with your customer demographic is recommended.
  • Barge-in handling — If a caller speaks while the agent is talking, the model may occasionally misinterpret the interruption. See Voice AI agent creation for barge-in configuration options.
  • Latency sensitivity — Voice conversations are real-time, so even small latency increases are noticeable. HoopAI optimizes for sub-second response times but network conditions can add delay.

Image and content limitations

  • Image generation style — AI-generated images may not perfectly match your brand style guide without detailed prompting.
  • Content length — Extremely long content pieces (over 3,000 words) may lose coherence toward the end. Breaking content into sections and generating each separately produces better results.
  • Factual claims in content — Always review AI-generated content for factual accuracy before publishing, especially for regulated industries.

Future model improvements and roadmap

HoopAI’s AI capabilities are evolving rapidly. Here is a look at what is on the horizon:
Support for image and document analysis within chat conversations, allowing customers to share photos or PDFs and receive AI-powered responses.
Expanded voice cloning and customization options, including the ability to fine-tune speaking style, pacing, and emotional tone for Voice AI agents.
More granular per-task model routing that considers the specific sub-task within a feature — for example, using a reasoning-optimized model for complex customer questions and a speed-optimized model for simple FAQs.
Broader language coverage with improved accuracy for non-English conversations, including real-time translation capabilities for Voice AI.
The ability to fine-tune models on your specific business data for even higher accuracy and brand consistency, available for enterprise plans.
Roadmap items are subject to change and do not represent commitments. Features will be announced through official release notes as they become available.

Next steps

Last modified on March 5, 2026