3. Security evaluation suites for agentic AI systems on AWS
Foundation models form the intelligence core of agentic AI systems. This makes their security characteristics critical to overall system safety. Systematic evaluation and testing of model behavior can help you identify vulnerabilities before deployment.
This section contains the following best practices:
3.1 Conduct model system card reviews (AI-specific)
Review model system cards
3.2 Use security evaluation suites (AI-specific)
Use model evaluation tools to probe your AI application with adversarial prompts that are designed to elicit security vulnerabilities or responsible AI failures. This systematic testing approach identifies potential vulnerabilities, such as:
-
Attempts to extract environment variables or credentials from the model
-
Prompts designed to generate malicious code or exploits
-
Information leakage of training data or sensitive information
Libraries and tools to support model evaluation include fmeval and SecEval