AI/automation ethics glossary
Security
AI/automation ethics glossary
Security
Security refers to the protection of an AI/automated system from breaches, leaks or unauthorised use in order to maintain the privacy and confidentiality of data and information - and to the capacity of AI itself to be weaponised as a tool of harm.
AI security as an ethical issue operates on two interlocking fronts.
The first is AI as a target, in which AI models and systems can be attacked, manipulated, or subverted in ways that traditional software cannot. Common vulnerabilities include:
Adversarial attacks: Crafting inputs that intentionally deceive AI models into misclassifying threats, allowing malware or intrusion attempts to slip through undetected.
Prompt injection: A technique where attackers manipulate the natural language inputs provided to AI systems, such as large language models, to override their original instructions or security controls, allowing attackers to make the AI reveal sensitive information, bypass authentication, or perform unauthorised actions.
Data poisoning: The act of intentionally sending false or misleading data inputs to influence a model's behaviour, typically with negative consequences. Types include targeted attacks, backdoor poisoning, label poisoning, and model inversion attacks.
The second front is AI as a weapon, where generative AI lowers the barrier to conducting sophisticated cyberattacks, disinformation campaigns, and social engineering at scale.
It also matters because insecure AI can spread misinformation, enable fraud, or support surveillance that undermines rights and democratic processes.
When AI security norms break down, the consequences cascade across multiple domains:
Physical harm. Poisoned datasets used in an autonomous car could prevent the recognition of certain traffic signs, causing accidents.
Institutional harm. Compromised AI systems used in hiring, policing, or benefits allocation can silently perpetuate injustice at scale.
Economic harm. Model theft, ransomware, and AI-enabled fraud impose substantial financial losses on individuals and organisations.
Erosion of trust. Poisoned AI systems can perpetuate biases, spread misinformation, and erode public trust in AI.
Democratic harm. AI-powered disinformation campaigns and deepfakes can undermine elections, public discourse, and institutional legitimacy.
AI and automation systems suffer from a variety of vulnerabilities, including:
Architectural vulnerability. AI models learn from data and are therefore fundamentally susceptible to manipulation of that data or of their inputs in ways traditional rule-based systems are not.
Speed of deployment. Commercial pressure to ship AI products quickly outpaces security testing and red-teaming.
Opacity and complexity. The black-box nature of many AI systems makes it difficult to detect or anticipate attack vectors.
Regulatory lag. Security standards for AI are nascent and unevenly applied across sectors and jurisdictions.
Open-source exposure. Widely available pre-trained models and datasets can contain embedded vulnerabilities or backdoors.
Expanded attack surface. As AI is integrated with third-party tools, APIs, browsers, and agents, the potential entry points for attack multiply rapidly.
Transparency versus security. Openness about AI architectures and training data is good for accountability and research, but also aids adversaries seeking to exploit vulnerabilities.
Access versus control. Open-source AI democratises innovation but also democratises access to powerful tools for malicious actors.
Offensive AI research. Security researchers must understand and replicate attacks to build defences, but doing so risks creating or disseminating dangerous capabilities.
Automation of security. Using AI to defend against AI-enabled threats may be less costly but introduces new failure modes and removes human judgment from consequential decisions.
Surveillance creep. Security monitoring of AI systems can justify expanded surveillance of users and populations, trading one harm for another.
Author: Charlie Pownall 🔗
Published: May 12, 2026
Last updated: May 12, 2026
You are welcome to use, copy, adapt, and redistribute this definition under a CC BY-SA 4.0 licence.
Let us know if you have any comments or suggestions about how to improve this definition, or would like to suggest and/or contribute additional terms to define.