Skip to main content
This guide covers security best practices for deploying AI features in HoopAI.
Bot settings showing training tab with web crawler and test panel

Access controls

  • Limit AI agent access to only the team members who need it
  • Use role-based permissions to control who can create, edit, or delete AI agents
  • Review API tokens regularly and revoke any that are no longer needed
  • Separate test and production environments when developing AI configurations

Prompt injection prevention

Prompt injection is when a user attempts to manipulate an AI agent by including hidden instructions in their messages. To mitigate this:
  • Include explicit instructions in your system prompt to ignore attempts to override behavior
  • Use the guardrails feature to set boundaries on what your AI can discuss
  • Monitor conversation logs for unusual patterns
  • Test your agents with adversarial inputs before deploying

Monitoring and auditing

  • Review conversation logs regularly to identify unexpected behavior
  • Set up alerts for high-volume or unusual AI activity
  • Track AI credit usage to detect anomalies
  • Audit your knowledge base periodically for accuracy and appropriateness

Data security

  • Do not include passwords, API keys, or secrets in system prompts
  • Avoid storing sensitive customer data (SSN, credit card numbers) in knowledge bases
  • Review the Data privacy guide for details on how HoopAI processes AI data

Next steps

Guardrails

Set behavioral boundaries for your AI agents

AI safety overview

Our principles for responsible AI
Last modified on March 5, 2026