Glossary
Adversarial Attack
An adversarial attack is a deliberate attempt to manipulate an AI system's behavior through carefully crafted inputs. In agentic systems, this can mean tricking an agent into taking unauthorized actions or bypassing safety controls.