Tag: LLM security

Threat Modeling for Large Language Model Integrations in Enterprise Apps 26 February 2026

Threat Modeling for Large Language Model Integrations in Enterprise Apps

Threat modeling for LLM integrations in enterprise apps is no longer optional. Learn the top five real-world risks-prompt injection, data poisoning, model theft, supply chain flaws, and insecure outputs-and how tools like AWS Threat Designer are making security practical for development teams.

Susannah Greenwood 6 Comments
Safety Layers in Generative AI: Content Filters, Classifiers, and Guardrails Explained 17 February 2026

Safety Layers in Generative AI: Content Filters, Classifiers, and Guardrails Explained

Safety layers in generative AI-like content filters, classifiers, and guardrails-are essential for preventing harmful outputs, blocking attacks, and protecting data. Without them, AI systems become unpredictable and dangerous.

Susannah Greenwood 10 Comments
Security and Compliance Considerations for Self-Hosting Large Language Models 28 January 2026

Security and Compliance Considerations for Self-Hosting Large Language Models

Self-hosting large language models gives organizations full control over data and compliance, but requires robust security, continuous monitoring, and deep expertise. Learn what it takes to do it right.

Susannah Greenwood 8 Comments
Training Data Poisoning Risks for Large Language Models and How to Mitigate Them 20 January 2026

Training Data Poisoning Risks for Large Language Models and How to Mitigate Them

Training data poisoning lets attackers corrupt AI models with tiny amounts of malicious data, causing hidden backdoors and dangerous outputs. Learn how it works, real-world examples, and proven ways to defend your models.

Susannah Greenwood 10 Comments
Prompt Injection Risks in Large Language Models: How Attacks Work and How to Stop Them 31 August 2025

Prompt Injection Risks in Large Language Models: How Attacks Work and How to Stop Them

Prompt injection attacks trick AI models into ignoring their rules, exposing sensitive data and enabling code execution. Learn how these attacks work, which systems are at risk, and what defenses actually work in 2025.

Susannah Greenwood 7 Comments