AI safety has become a cybersecurity issue because modern AI systems are now targets, tools, and attack surfaces. They can be attacked directly through prompt injection, data poisoning, model extraction, and supply-chain compromise. They can also be used to commit fraud, phishing, impersonation, reconnaissance, and code exploitation on a larger scale.
It is precisely why current NIST guidance now treats secure AI, AI-enabled defense, and AI-enabled attacks as distinct but intertwined cybersecurity problems. According to the NIST Artificial Intelligence Risk Management Framework, AI carries “risks that are not comprehensively addressed by current risk frameworks and approaches.”
Reuters documented AI deepfakes in the 2026 U.S. midterms and showed how consumer-facing chatbots could assist realistic phishing campaigns. The news cycle is telling a consistent story: the lived risk is not only unsafe output but also fraud, access compromise, operational misuse, and degraded trust in evidence itself. The consequence is that organizations should stop treating AI as a mere productivity layer and start treating it as a cyber system that needs identity controls, software assurance, logging, procurement scrutiny, model evaluation, human approval gates, and incident response planning.
Why AI safety is a cybersecurity issue
When an AI system interacts with sensitive data, identity, code, operational controls, external tools, or human decision-making, AI safety becomes a cybersecurity issue. NIST's AI framework says that AI is different from regular software because its models, data, and contexts of use are always changing, hidden, and part of society. The framework says that AI systems can change over time, fail in ways that aren't obvious, and cause problems that regular software controls do not always see coming.
The above-mentioned study notes, “AI systems and the contexts in which they are deployed are frequently complex.” NIST’s generative-AI profile makes the cyber link explicit. It says that generative AI makes information security, data privacy, phishing, malicious code generation, model abuse, and value-chain exposure risks worse or new. AI safety is about whether a chatbot says something harmful or false, whether a model can be changed, whether it can change users, and whether an attacker can use it.
How AI safety appears in the current news cycle
According to the FBI Internet Crime Complaint Center, “AI-enabled synthetic content is becoming increasingly difficult to detect and easier to make.” The dominant news pattern is practical, not philosophical. In May 2025, the FBI warned that attackers were using text and AI-generated voice messages to impersonate senior U.S. officials and gain access to personal accounts.
In September 2025, Reuters showed that mainstream chatbots could help plan persuasive phishing campaigns. In December 2025, CISA and partner agencies released a joint guide on secure AI integration in operational technology. In March 2026, Reuters reported that AI-generated deepfake campaign ads had entered the U.S. midterms in a landscape with “no federal regulation” governing AI use in political messaging. In April 2026, the FBI’s annual cybercrime report quantified AI-related scam losses at more than $893 million.
The stories cluster around four themes.
- Identity deception: voice cloning, fabricated images, deepfake videos, and fake profiles.
- Persuasive scale: AI lowers the labor cost of phishing, romance scams, investment fraud, and business email compromise.
- Operational integration: the CISA OT guidance reflects the fact that AI is moving into environments where mistakes can have physical, safety, and continuity consequences.
- Governance lag: election messaging, procurement, and embedded vendor use of AI are moving faster than uniform federal controls.
What is AI related cyber risk?
The most useful working definition is the group of cybersecurity risks that AI systems and workflows using AI create, change, make worse, or hide. It encompasses risks to AI systems, risks from AI systems, risks through AI systems, and risks surrounding AI systems, including supply-chain and procurement vulnerabilities. NIST's adversarial-machine-learning taxonomy is especially useful here because it looks at risk throughout the whole machine’s lifecycle, not just when it is deployed.
The NIST's generative-AI profile says that prompt injection means changing what input is given to a generative AI system so that it acts in ways that are not intended. The definition includes a number of specific types of risk. There is a risk to both the model and the data, including poisoning, extraction, membership inference, leaking training or retrieved content, and privacy harm. In a high-trust process, there is a risk that a model-making tool will make calls that lead to insecure code, misclassify an event, or create false evidence.
There is a risk of adversarial use, where attackers use AI to improve how they choose their targets, write malicious lures, or automate parts of the intrusion chain. An ecosystem risk also exists, which comes from model providers, libraries, APIs, cloud dependencies, fine-tuned components, and third-party data pipelines that are passed down from one generation to the next.
The difference between AI being used to assist defenders and AI being used to assist attackers
Defenders usually do better when there is a lot of pattern recognition, classification, summarization, triage, or finding strange things. Attackers tend to do better when labor, language quality, personalization, and early-stage reconnaissance have historically limited growth. The Defend and Thwart focus areas in NIST's Cyber AI Profile make this split official. Studies also show that attackers seem to have the most success in the early kill-chain phases.
The paper, Impact of AI on the Cyber Kill Chain: A Systematic Review states, "Our findings indicate that AI-based tools are used most effectively in the initial stages of cyberattacks.” The main difference is that defenders need reliable integration, while attackers only need output that is good enough. A Security Operations Center copilot that is wrong too often makes analysts work more slowly. An email made by a scammer's AI only needs to fool a small number of people. Because of this, fraud and social engineering are more likely to show near-term AI risk than fully autonomous, movie-style attacks.
Why does limited federal regulation make the cybersecurity side of AI harder to manage?
The U.S. problem is not regulatory absence in an absolute sense; it is fragmentation. There are federal laws, sectoral obligations, agency memoranda, enforcement tools, procurement rules, and voluntary standards. What is missing is a single comprehensive legal framework that cleanly governs AI, privacy, security, accountability, and cross-sector deployment together. The result is uneven obligations, uneven visibility, and uneven incentives.
The report AI OMB Action Needed to Address Privacy Related Gaps in Federal Guidance states, “There is not a comprehensive legal framework governing AI, privacy, or their intersection.” The report noted that the 2023 executive order on safe and trustworthy AI was rescinded in January 2025, and the White House then issued an order focused on removing barriers to American AI leadership. The Office of Management and Budget followed in April 2025 with M-25-21 and M-25-22, which still preserve governance and acquisition guidance for federal use, but they do so in a policy environment that is more innovation-forward and less centered on a single, cross-government safety architecture.
See also: HIPAA Compliant Email: The Definitive Guide (2026 Update)
FAQs
Is AI safety just about chatbots giving wrong answers?
No. Wrong answers are only part of the problem. AI safety also covers prompt injection, data leakage, model misuse, insecure integrations, fake identities, and over-trust in outputs that sound confident but are wrong.
What makes AI systems harder to secure than traditional software?
AI systems are harder to secure because they depend on models, data, prompts, context, integrations, and human interpretation rather than fixed logic alone.
How can AI increase risk without being openly malicious?
AI can increase risk simply by being wrong, overconfident, poorly integrated, or given too much access.
Subscribe to Paubox Weekly
Every Friday we bring you the most important news from Paubox. Our aim is to make you smarter, faster.
