TRUST & SECURITY IN GENERATIVE AI
Our Testers and AI Engineers work collaboratively with our Client's team including ethicists and domain specialists to build and maintain responsible and trustworthy Generative AI systems.
OWASP TOP 10 FOR LLM RISKS
The OWASP Top 10 for Large Language Model (LLM) Risks provides a framework for identifying and mitigating security risks associated with the use of LLMs. Below are the best practices aligned with these risks:
LLM01. PROMPT INJECTIONRisk: Malicious input designed to manipulate the LLM into executing unintended actions.
Prevention & Mitigation Strategies:
LLM02. SENSITIVE INFORMATION DISCLOSURE
Risk: Sensitive or proprietary data unintentionally exposed via model outputs (data leakage).
Prevention & Mitigation Strategies:
LLM03. SUPPLY CHAIN
Risk: Unsafe outputs such as injection attacks (e.g., SQL injection, XSS).
Prevention & Mitigation Strategies:
LLM04. DATA & MODEL POISONING
Risk: Adversaries insert malicious data into training datasets.
Prevention & Mitigation Strategies:
LLM05. IMPROPER OUTPUT HANDLING
Risk: Unauthorized access or theft of proprietary models.
Prevention & Mitigation Strategies:
LLM06. EXCESSIVE AGENCY
Risk: Abuse of LLM capabilities by unauthorized users.
Prevention & Mitigation Strategies:
LLM07. SYSTEM PROMPT WEAKNESS
Risk: Adversarially crafted inputs that exploit model weaknesses.
Prevention & Mitigation Strategies:
LLM08. VECTOR & EMBEDDING WEAKNESS
Risk: Blind trust in model outputs leading to inaccurate or harmful actions.
Prevention & Mitigation Strategies:
LLM09. MODEL MISUSE
Risk: Using LLMs for malicious or unintended purposes (e.g., generating disinformation).
Prevention & Mitigation Strategies::
LLM10. UNBOUNDED CONSUMPTION
Risk: A Large Language Model (LLM) application allows users to conduct excessive and uncontrolled inferences, leading to risks
such as denial of service (DoS), economic losses, model theft, and service degradation.
Prevention & Mitigation Strategies:
The OWASP Top 10 for Large Language Model Applications Project aims to educate developers, designers, architects, managers, and organizations about the potential security risks when deploying and managing Large Language Models (LLMs) and Generative AI applications. The project provides a range of resources. Most notably the OWASP Top 10 list for LLM applications listing the top 10 most critical vulnerabilities often seen in LLM applications, highlighting their potential impact, ease of exploitation, and prevalence in real-world applications. Visit OWASP.org
RED TEAMING & EVALUATIONS
Red-teaming in the context of generative AI refers to the practice of rigorously testing and probing AI models to identify vulnerabilities, biases, risks, and potential misuse cases. The goal is to improve the safety, robustness, and ethical performance of these systems by simulating adversarial or challenging scenarios.
1. Adversarial Testing:
2. Bias and Fairness Analysis:
3. Safety and Harm Prevention:
4. Misuse Scenarios:
5. Robustness Testing:
6.Iterative Feedback and Improvement:
Why Is Red-Teaming Important?
Red Teaming is often performed by dedicated teams of experts, which may include ethicists, domain specialists, adversarial testers, and AI engineers. The insights gained are critical for building responsible and trustworthy AI systems.
GUARDRAILS ON AMAZON BEDROCK
With Amazon Bedrock, you can: