AICM AtlasCSA AI Controls Matrix
All threats
Threat Lens

Model manipulation

This category involves attempts to evade detection or manipulate the LLM model to produce inaccurate or misleading results. It encompasses techniques, such as direct or indirect prompt injection (adversarial inputs), which aim to exploit vulnerabilities in the model's understanding and decision-making processes.

114 controls mitigate this threat across 15 domains

DomainsControls