Disclaimer: this is a report generated with my tool: https://github.com/DTeam-Top/tsw-cli. See it as an experiment not a formal research, 😄。


Mindmap

Mindmap

Summary

This document outlines the OWASP Top 10 vulnerabilities for Large Language Model (LLM) applications in 2025. It serves as a guide for developers, security professionals, and users to understand and mitigate the most critical risks associated with LLMs, including prompt injection, sensitive information disclosure, supply chain vulnerabilities, and misinformation. The document provides descriptions of each vulnerability, common examples, prevention strategies, attack scenarios, and references to related frameworks and taxonomies.

Terminology

  • LLM (Large Language Model): A deep learning model designed to understand and generate human-like text based on patterns learned from vast amounts of data.
  • Prompt Injection: Manipulating an LLM's input to alter its behavior or output in unintended ways.
  • RAG (Retrieval-Augmented Generation): A technique that enhances LLM responses by retrieving information from external knowledge sources.
  • SBOM (Software Bill of Materials): A comprehensive inventory of components used in software, aiding in vulnerability management.
  • LoRA (Low-Rank Adaptation): A fine-tuning technique that efficiently adapts pre-trained LLMs to specific tasks.
  • PEFT (Parameter-Efficient Fine-Tuning): A set of techniques for fine-tuning LLMs with minimal computational resources.
  • PII (Personally Identifiable Information): Any data that could potentially identify a specific individual.
  • XSS (Cross-Site Scripting): A type of web security vulnerability that allows attackers to inject malicious scripts into websites viewed by other users.
  • CSRF (Cross-Site Request Forgery): A type of web security vulnerability that allows attackers to induce users to perform actions they do not intend to perform.
  • SSRF (Server-Side Request Forgery): A type of web security vulnerability that allows an attacker to cause the server-side application to make HTTP requests to an arbitrary domain of the attacker's choosing.
  • API (Application Programming Interface): A set of rules and specifications that software programs can follow to communicate with each other.

Main Points

Point 1: LLM01:2025 Prompt Injection

Prompt injection vulnerabilities arise when user inputs unintentionally or maliciously alter the LLM's behavior. This can lead to harmful content generation, unauthorized access, or manipulation of critical decisions.

Implementation:

  • Direct Prompt Injections: User input directly alters the model's behavior.
  • Indirect Prompt Injections: External sources contain data that, when interpreted by the model, alters its behavior.

Mitigation:

  • Constrain model behavior with specific instructions.
  • Validate expected output formats.
  • Implement input and output filtering.
  • Enforce privilege control.
  • Require human approval for high-risk actions.
  • Segregate external content.
  • Conduct adversarial testing.

Point 2: LLM02:2025 Sensitive Information Disclosure

LLMs can inadvertently expose sensitive data, including PII, financial details, and confidential business data, leading to unauthorized access and privacy violations.

Mitigation:

  • Integrate data sanitization techniques.
  • Enforce strict access controls.
  • Utilize federated learning and differential privacy.
  • Educate users on safe LLM usage.
  • Ensure transparency in data usage.
  • Conceal system preambles.

Point 3: LLM03:2025 Supply Chain

LLM supply chains are vulnerable to compromised training data, models, and deployment platforms, resulting in biased outputs and security breaches.

Implementation:

  • Managing third-party risks like outdated components and licensing issues.
  • Mitigating vulnerabilities in pre-trained models and LoRA adapters.
  • Addressing risks in collaborative development environments and on-device LLMs.

Mitigation:

  • Vet data sources and suppliers.
  • Apply vulnerability scanning and management.
  • Conduct AI red teaming and evaluations.
  • Maintain a Software Bill of Materials (SBOM).
  • Use verifiable model sources and integrity checks.
  • Implement monitoring and auditing for collaborative model development.

Point 4: LLM04: Data and Model Poisoning

Data poisoning involves manipulating training data to introduce vulnerabilities or biases, compromising model security and performance.

Mitigation:

  • Track data origins and transformations.
  • Vet data vendors rigorously.
  • Implement strict sandboxing.
  • Tailor models for specific use cases.
  • Ensure sufficient infrastructure controls.
  • Use data version control (DVC).
  • Test model robustness with red team campaigns.
  • Monitor training loss and analyze model behavior.

Point 5: LLM05:2025 Improper Output Handling

Insufficient validation and sanitization of LLM outputs can lead to XSS, CSRF, SSRF, privilege escalation, and remote code execution.

Mitigation:

  • Treat the model as any other user with proper input validation.
  • Follow OWASP ASVS guidelines.
  • Encode model output back to users.
  • Implement context-aware output encoding.
  • Use parameterized queries for database operations.
  • Employ strict Content Security Policies (CSP).
  • Implement robust logging and monitoring systems.

Point 6: LLM06:2025 Excessive Agency

Granting LLMs excessive functionality, permissions, or autonomy can lead to damaging actions due to unexpected or manipulated outputs.

Mitigation:

  • Minimize extensions and their functionality.
  • Avoid open-ended extensions.
  • Minimize extension permissions.
  • Execute extensions in the user's context.
  • Require user approval for high-impact actions.
  • Implement authorization in downstream systems.
  • Sanitize LLM inputs and outputs.

Point 7: LLM07:2025 System Prompt Leakage

System prompt leakage exposes sensitive information used to guide the LLM's behavior, potentially facilitating other attacks.

Mitigation:

  • Separate sensitive data from system prompts.
  • Avoid reliance on system prompts for strict behavior control.
  • Implement guardrails outside of the LLM.
  • Enforce security controls independently from the LLM.

Point 8: LLM08:2025 Vector and Embedding Weaknesses

Weaknesses in vector and embedding generation, storage, or retrieval can lead to harmful content injection, manipulated outputs, or sensitive information access.

Mitigation:

  • Implement fine-grained access controls and permission-aware vector stores.
  • Implement robust data validation pipelines.
  • Review data combinations and classifications.
  • Maintain detailed logs of retrieval activities.

Point 9: LLM09:2025 Misinformation

LLMs can produce false or misleading information that appears credible, leading to security breaches, reputational damage, and legal liability.

Mitigation:

  • Use Retrieval-Augmented Generation (RAG).
  • Enhance the model with fine-tuning or embeddings.
  • Encourage cross-verification and human oversight.
  • Implement automatic validation mechanisms.
  • Communicate risks to users.
  • Establish secure coding practices.

Point 10: LLM10:2025 Unbounded Consumption

Uncontrolled LLM inferences can lead to denial of service (DoS), economic losses, model theft, and service degradation.

Mitigation:

  • Implement strict input validation.
  • Limit exposure of logits and logprobs.
  • Apply rate limiting and user quotas.
  • Manage resource allocation dynamically.
  • Set timeouts and throttling.
  • Restrict the LLM's access using sandboxing.
  • Monitor resource usage and detect anomalies.

Improvements And Creativity

  • Focus on Real-World Risks: The 2025 update reflects a deeper understanding of existing risks and introduces critical updates based on real-world applications.
  • New Vulnerabilities: The inclusion of "System Prompt Leakage" and "Vector and Embeddings Weaknesses" addresses previously overlooked areas with significant exploit potential.
  • Expanded Scope: "Unbounded Consumption" expands beyond Denial of Service to include resource management and cost concerns.
  • Emphasis on Agentic Architectures: The "Excessive Agency" entry is expanded to address the increasing use of LLMs in autonomous agent systems.

Insights

This document provides a timely and essential resource for securing LLM applications. As LLMs become more integrated into various systems, understanding and mitigating these vulnerabilities is crucial.

Predictions and Recommendations:

  • Security measures should focus on a defense-in-depth strategy, combining multiple mitigation techniques.
  • Continuous monitoring and adaptation are necessary due to the evolving nature of LLM vulnerabilities.
  • Further research is needed to develop robust defenses against advanced attacks, such as multimodal prompt injection and model extraction.

References

Original Paper: OWASP Top 10 for Large Language Models (LLM) Applications - 2025


Report generated by TSW-X
Advanced Research Systems Division
Date: 2025-03-17 09:37:18.887920