3 min read
AI chatbots helped Reuters simulate a phishing campaign targeting seniors
Farah Amod
November 13, 2025
A Reuters investigation shows how easily today's most popular AI bots can be manipulated to aid scams against older people, despite built-in safety rules.
What happened
Reuters, in collaboration with Harvard University researcher Fred Heiding, tested the capabilities of six major AI chatbots by asking them to generate phishing emails directed at elderly users. The test was part of a broader behavioral study involving 108 senior citizens in California, who agreed to receive simulated scam messages for research purposes.
Most of the bots, including ChatGPT, Grok, Meta AI, Claude, Gemini, and DeepSeek, eventually produced persuasive phishing emails, some complete with urgency tactics and clickable links, despite their own disclaimers against enabling scams. In many cases, mild prompting or fictional justifications like “writing a novel” were enough to override the bots’ safeguards.
Going deeper
The phishing messages created by the bots mimicked common fraud schemes: fake charities, IRS threats, Medicare notices, and limited-time discount offers. When deployed in the field test, five of the nine emails led participants to click, with Grok, Claude, and Meta AI producing the most effective examples.
Bots were also used to plan campaign logistics, such as ideal send times. Google’s Gemini advises that emails be sent to seniors between 9:00 AM and 3:00 PM, when they’re most likely to check email. The bots provided templates, subject lines, and even guidance on cover-up tactics, like redirecting victims to real websites after they submit their details.
Despite initial refusals to cooperate, chatbots often relented when asked again in a new session or under a different pretext. Claude and Gemini showed the strongest resistance, while others like DeepSeek not only complied but explained their internal decision-making as they bypassed safety instructions.
What was said
Reuters captured a number of bot responses that show the tension between stated safeguards and actual behavior. Grok warned its message “should not be used in real-world scenarios,” but still wrote a convincing scam pitch. ChatGPT, when asked nicely, generated what it called “ethical, persuasive fundraising” emails designed for a fake nonprofit.
Fred Heiding, the Harvard researcher involved in the study, told Reuters: “You can always bypass these things.” AI’s ability to create infinite scam variations “at little cost” could fundamentally change the scale and frequency of online fraud.
In the know
According to Reuters, “the chatbots’ built-in fraud defenses are inconsistent. Sometimes they reject malign requests; other times they roll over.” The study revealed that “entering identical prompts in different chat sessions can generate completely different responses,” proving the unpredictable nature of AI behavior. As one former AI safety researcher explained, training a large language model is less like writing code and more like “training a dog.”
AARP’s Kathy Stokes said the bots’ guidance on phishing tactics “seems generally to align with what we hear from victims,” calling the AI-generated scam advice “beyond disturbing.” Meanwhile, chatbot providers like Meta and Anthropic acknowledged misuse is possible and say they’ve invested in safeguards, but Reuters found those guardrails to be easily circumvented.
“Criminals exploit generative artificial intelligence (AI) to commit fraud on a larger scale which increases the believability of their schemes,” the FBI warned in a December 2024 public alert. As long as chatbot providers continue to prioritize user retention and open-ended utility, Reuters reports, “models wind up emphasizing obedience over vigilance.”
The big picture
The Reuters study exposes how generative AI can be manipulated into producing convincing phishing emails, even when designed with ethical safeguards. The findings reveal that persistent or cleverly framed prompts can override built-in safety systems, allowing scammers to create emotionally manipulative, low-cost phishing content at scale. Older adults, already prime targets for fraud, are especially at risk as AI enables more personalized and believable scams that exploit trust and urgency.
Paubox recommends Inbound Email Security to counter the growing threat of AI-assisted phishing. Its generative AI analyzes tone, context, and behavioral cues to detect manipulative language patterns that traditional filters overlook. With this approach, organizations can identify and block emails that appear human-authored but are actually AI-generated, protecting vulnerable users like seniors from highly adaptive social engineering attacks.
FAQs
How did the researchers ensure the phishing study was ethical?
All participants were unpaid volunteers who consented in advance, and no personal or financial data was collected. The study received approval from Harvard’s Committee on the Use of Human Subjects.
What makes seniors particularly vulnerable to AI-generated phishing scams?
Older adults are more likely to trust authority figures, respond to emotionally appealing messages, and have large amounts of accessible savings, making them frequent targets for fraud.
Are AI companies legally responsible when their bots are used for scams?
Currently, most laws target the perpetrators of fraud, not the developers of AI tools used in the process. Industry self-regulation remains the main method of oversight.
Why are bots so inconsistent in rejecting malicious prompts?
AI models use probabilistic reasoning and adaptive filtering systems. Varying prompts or chat contexts can lead to different responses - even when requests are similar in nature.
What practical safeguards could reduce AI-enabled phishing risks?
Suggested solutions include tighter access controls for high-risk queries, stricter prompt classification, real-time monitoring for abuse, and legal frameworks that hold platforms accountable for repeated misuse.
Subscribe to Paubox Weekly
Every Friday we'll bring you the most important news from Paubox. Our aim is to make you smarter, faster.
