Creating and Detecting Phishing Emails Using Large Language Models
This research demonstrates how large language models (LLMs) can be leveraged to generate highly convincing, personalized phishing emails with minimal effort. The study introduces the 'V-Triad' framework, which combines credibility, relevancy, and customizability to create sophisticated social engineering lures. The researchers evaluate the effectiveness of these AI-generated emails against human targets and discuss how LLMs can be repurposed for defensive purposes, such as improved spam detection and personalized cybersecurity training. The talk highlights the shift in the phishing landscape toward automated, high-quality content generation.
Why Your Phishing Simulations Are Already Obsolete
TLDR: Large language models have fundamentally shifted the economics of social engineering by enabling the rapid generation of highly personalized, context-aware phishing lures. Research presented at Black Hat 2023 demonstrates that AI-generated emails consistently outperform traditional, generic templates in user engagement. Security teams must move beyond simple link-clicking metrics and start testing for cognitive resilience against sophisticated, AI-augmented deception.
The era of the "Nigerian Prince" email is effectively over. For years, security awareness training has relied on teaching employees to spot obvious red flags: poor grammar, mismatched sender domains, and generic greetings. These indicators were the bread and butter of phishing detection. Today, that baseline is useless. Large language models like GPT-4, Claude, and Llama have lowered the barrier to entry for creating high-quality, contextually relevant lures that bypass traditional human heuristics.
The V-Triad Framework for Automated Deception
Research from Harvard University, presented at Black Hat 2023, introduces the V-Triad framework to quantify why AI-generated phishing is so effective. The framework breaks down a successful lure into three components: credibility, relevancy, and customizability.
Credibility is the gatekeeper. It is the visual and linguistic polish that makes an email look legitimate. Historically, this required significant manual effort from an attacker. Now, an LLM can generate a professional, error-free email in seconds. Relevancy is the hook. It connects the email to the target’s specific environment, such as mentioning a recent campus event or a specific company policy. Customizability is the final piece, allowing an attacker to iterate on the message based on the target's response or specific profile.
When you combine these three, you get a lure that doesn't just look real—it feels real. The researchers found that by feeding minimal data points into an LLM, they could generate emails that were indistinguishable from legitimate corporate communications. The technical barrier to creating these lures is now effectively zero. If you can write a prompt, you can launch a campaign.
The Mechanics of the Attack
The attack flow is straightforward. An attacker collects background information on a target—which is trivial given the massive digital footprint most professionals leave on platforms like LinkedIn or Twitter. Once the attacker has a few data points, they use an LLM to draft the email.
Prompt: Create an email for Harvard students about a summer update in the
Harvard Shuttle's operating hours, and provide a link to the updated
timetable.
The output is a polished, professional email. The researchers noted that while some models have safety guardrails to prevent the generation of malicious content, these are easily bypassed. A simple persona shift—telling the model you are a researcher conducting a study—is often enough to strip away those protections.
The most dangerous aspect of this technique is the "unsubscribe" link. Many security awareness programs teach users to click "unsubscribe" if they suspect an email is spam. The researchers found that this is a powerful phishing vector. By including a legitimate-looking unsubscribe link, attackers can confirm that an email address is active and that the target is willing to interact with the message. This is a classic T1566 phishing technique, now supercharged by AI.
Real-World Applicability for Pentesters
For those of us conducting red team engagements, this changes the game. We no longer need to spend hours crafting the perfect lure. We can automate the entire process, from reconnaissance to content generation. During an engagement, you can now generate hundreds of unique, highly personalized emails in the time it used to take to write one.
The impact is clear: higher click-through rates and a significantly higher likelihood of credential harvesting. If you are still using static, generic phishing templates for your internal testing, you are not accurately measuring your organization's risk. You are testing their ability to spot 2010-era spam, not their ability to defend against modern, AI-augmented threats.
Shifting to Cognitive Resilience
Defenders need to stop focusing on the "what" and start focusing on the "why." If an email looks perfect, the traditional red flags are gone. The only remaining indicator is the intent. Does this email make sense in the context of the recipient's daily workflow?
Organizations should prioritize training that emphasizes cognitive pauses. The research suggests that the most effective defense is simply slowing down. If an email creates an artificial sense of urgency, the recipient should be trained to verify the request through an out-of-band channel, such as a direct message to the supposed sender or a check of the official company portal.
We also need to leverage these same models for defense. LLMs are excellent at analyzing intent. By integrating an LLM into your email security stack, you can move beyond simple keyword filtering. You can analyze the semantic intent of an email and flag messages that, while grammatically perfect, exhibit the psychological hallmarks of a social engineering attempt.
The landscape of social engineering has shifted. The tools that once required a skilled human operator are now available to anyone with an internet connection. If you are not accounting for this in your security strategy, you are already behind. Stop looking for typos and start looking for the manipulation of your team's decision-making process. The next time you receive an email that seems just a little too helpful, take a breath. Your rational brain is your best defense.
Vulnerability Classes
Target Technologies
Attack Techniques
Up Next From This Conference

Chained to Hit: Discovering New Vectors to Gain Remote and Root Access in SAP Enterprise Software

Zero-Touch-Pwn: Abusing Zoom's Zero Touch Provisioning for Remote Attacks on Desk Phones

ODDFuzz: Hunting Java Deserialization Gadget Chains via Structure-Aware Directed Greybox Fuzzing
Similar Talks

Inside the FBI's Secret Encrypted Phone Company 'Anom'

Kill List: Hacking an Assassination Site on the Dark Web

