porkeynote
Menu
  • Home
  • About
  • Categories
    • Urban Fiction
    • People
    • CyberSecurity – My Journey
Menu

AI Is Now Fighting AI: What This Means for Cybersecurity

Posted on March 23, 2026 by ndiki

We are no longer in the era of:

  • Hackers manually writing phishing emails
  • Security teams manually reviewing alerts
  • Attacks unfolding over days and weeks

We are now in the era of:

  • AI writing perfect phishing campaigns in seconds
  • AI defending against attacks in milliseconds
  • Battles fought at speeds humans cannot perceive

AI is now fighting AI. And everything has changed.

The attacker’s arsenal: how AI is weaponized

The numbers are staggering

87% of organizations worldwide experienced AI-driven cyberattacks in 2025.

73% of security professionals say AI-powered threats are already hitting their organizations.

AI-powered cyberattacks have increased by 72% year-over-year globally.

This isn’t theoretical. This is operational.

What AI enables attackers to do

1. Hyper-personalized phishing at scale

82.6% of phishing emails detected between September 2024 and February 2025 utilized AI – a 53.5% year-on-year increase.

Security teams report a staggering 1,265% surge in phishing attacks linked to generative AI since 2023.

Traditional phishing relied on volume: send millions of poorly written emails and hope a few people clicked.

  • AI changes the equation entirely:
  • Flawless grammar and tone – No more spelling errors to spot
  • Perfect impersonation – Mimics CEO writing style exactly
  • Contextual awareness – References recent company events, projects, relationships
  • Cultural adaptation – Appropriate formality, humor, regional expressions
  • Timing optimization – Sent when targets are most likely to respond
  • Endless variation – Each email unique to evade filters
  • A/B testing at scale – Tries different approaches, learns what works

Real example:

A 2024 attack against a mid-sized healthcare organization used AI to scrape employee LinkedIn profiles, identifying 47 staff members who had recently completed cybersecurity certifications.

The attackers sent personalized “certificate verification” phishing emails that achieved a 38% click rate by exploiting the recency of the legitimate activity.

In one academic study, AI-crafted phishing emails achieved 54% click rates, compared to 12% for human-written ones.

2. Deepfake voice and video fraud

85% of organizations reported some form of deepfake attack in 2025.

Deepfake incidents grew 19% in Q1 2025 compared to all of 2024.

Real Case: $25 million lost in single deepfake incident – Hong Kong financial firm fell victim to sophisticated deepfake video conference call impersonating CFO.

How it works:

Attackers use AI to:

  • Clone executive voices from publicly available audio (earnings calls, interviews, podcasts)
  • Generate deepfake video for “live” video calls
  • Mimic speech patterns, mannerisms, and decision-making style
  • Conduct real-time video conferences with convincing deepfakes

The victim doesn’t stand a chance without AI-powered detection.

3. Autonomous attack chains

This is where it gets terrifying.

AI enables fully autonomous attacks that identify vulnerabilities, craft exploits, and launch multi-stage campaigns with little to no human input.

Alleged cases of large-scale cyber-espionage have been executed with minimal human involvement, with AI handling reconnaissance, initial access, privilege escalation, and exfiltration in one coordinated chain.

to test what happens when threat actors harness agentic AI:

The results:

  • Automated reconnaissance – AI scans networks, identifies targets, maps relationships
  • Dynamic exploit generation – Creates custom exploits for discovered vulnerabilities
  • Real-time adaptation – Changes tactics when detection systems activate
  • Multi-stage coordination – Executes privilege escalation → lateral movement → data exfiltration seamlessly
  • Deception at scale – Manipulates threat indicators to confuse analysts
  • TTP mimicry – Impersonates known threat actors to obscure attribution

Speed: Minutes, not weeks.

4. Polymorphic Malware

92% of polymorphic phishing attacks utilize AI.

76.4% of all phishing attacks in 2024 had at least one polymorphic feature.

What is polymorphic malware?

Malware that changes its code signature with each deployment while maintaining the same function.

Traditional antivirus: Blocks known malware signatures Polymorphic malware: Different signature every time Result: Traditional defenses useless

AI enables:

  • Automatic code mutation
  • Evasion of signature-based detection
  • Real-time adaptation to security responses
  • Benign-looking code that confuses scanners

AI-generated content was used to obfuscate ransomware payloads by filling scripts with benign randomly generated text to confuse security scanners.

5. Credential Theft and Business Email Compromise (BEC)

BEC was responsible for $2.77 billion in reported losses in the US in 2024.

Since 2015, total BEC losses have exceeded $17.1 billion over the past decade—an increase of over 1,025%.

The average amount requested in wire transfer BEC attacks in Q4 2024 was $128,980.

AI makes BEC devastatingly effective:

  • Scrapes public data for target identification
  • Analyzes email patterns to perfect impersonation
  • Times attacks for maximum success (end of quarter, Friday afternoons, holidays)
  • Generates invoices that match vendor formatting exactly
  • Creates urgency without raising suspicion

6. Zero-Day Exploitation

41% of zero-day vulnerabilities in 2025 were discovered through AI-assisted reverse engineering by attackers.

AI can:

  • Analyze software for vulnerabilities faster than human researchers
  • Generate exploits automatically
  • Test exploits at scale
  • Weaponize vulnerabilities before vendors patch them

The window between discovery and exploit: Hours, not weeks.

The defender’s response: how AI fights back

But defenders aren’t sitting idle.

51% of enterprises now use security AI or automation, and those organizations experience $1.8 million lower average breach costs than those without it.

Companies using AI-driven security platforms report detecting threats up to 60% faster than those using traditional methods.

What AI enables defenders to do

1. Real-Time Threat Detection

AI-augmented SOCs have demonstrated a 50% reduction in mean time to detect (MTTD).

How it works:

Instead of relying on predefined rules or known patterns, defensive AI understands context:

  • How real employees communicate
  • How vendor relationships normally function
  • What typical workflows look like
  • What “normal” looks like for each user

AI can identify subtle deviations that indicate fraud—even when the email itself looks flawless.

Example:

AI flags an email because:

  • Writing matches CEO style too perfectly (humans vary; AI doesn’t)
  • Request timing suspicious (sent 47 seconds after calendar showed “Available”)
  • Linguistic patterns show micro-inconsistencies invisible to humans
  • Metadata doesn’t match historical communication patterns

Traditional security: Misses it AI security: Blocks it

2. Behavioral Anomaly Detection

Defensive AI baselines normal user and service account behavior:

  • Login patterns
  • Device posture
  • Token usage
  • Access frequency

It flags anomalous activity in real time:

  • Impossible travel (login from New York 10 minutes after login from London)
  • Suspicious OAuth consent grants
  • Session hijacking indicators
  • Privilege escalation attempts

Speed: Milliseconds

3. Automated Incident Response

When a threat is detected, AI can:

  • Isolate compromised hosts
  • Block malicious IPs/domains
  • Reset credentials
  • Contain threats before they spread
  • Initiate remediation workflows

All automatically. No human required.

Post-incident, these systems provide intelligence that strengthens defenses against future attacks.

4. Predictive Threat Intelligence

Agentic AI doesn’t just react. It predicts.

Agentic AI is the next generation of modern threat intelligence, giving defenders the speed and autonomy attackers already exploit.

Instead of reacting to threats, Agentic AI predicts and responds across the full attack lifecycle.

Unlike conventional AI, Agentic AI evaluates scenarios, prioritizes risks, and initiates responses with human-like judgment at machine speed.

5. Continuous Learning and Adaptation

Every attack teaches defensive AI:

  • New attack patterns
  • Evasion techniques
  • Emerging threats
  • Adversary tactics

The AI gets smarter with every encounter.

Who’s winning the arms race?

The answer is: It depends.

98% of security leaders say AI has already been widely used in cyberattacks against their organizations.

But organizations using AI-driven security platforms typically achieve around 95% detection accuracy versus 85% with traditional tools.

They also reduce breach costs by an average of $1.9 million.

The divide is forming:

Organizations WITH AI defense:

  • Detect threats 60% faster
  • Contain breaches 108 days faster
  • Save $1.8-$2.2 million per breach
  • Stop attacks before damage occurs

Organizations WITHOUT AI defense:

  • Get overwhelmed by volume
  • Miss sophisticated attacks
  • Respond in hours/days (too slow)
  • Suffer catastrophic losses

As one expert notes: “76% of organizations cannot match AI attack speed, creating a pivotal period where offensive AI may temporarily outpace defenses.”

We’re in that window right now.

What this means for organizations

If you’re not using AI for security in 2026, you’re vulnerable. Period.

Here’s why:

1. Traditional defenses are obsolete

Signature-based detection? Useless against polymorphic malware.

Rule-based filters? Bypassed by AI-generated content.

Human analysis? Too slow by orders of magnitude.

You need AI to fight AI.

2. The window is closing

76% of organizations cannot match AI attack speed.

This creates a dangerous window where attackers have the advantage.

Every month you delay AI adoption, you fall further behind.

3. Partial solutions don’t work

You can’t just add “some AI” to your security stack.

You need:

  • AI-powered threat detection
  • AI-driven incident response
  • AI-enabled behavioral analytics
  • Agentic AI for predictive defense
  • Human oversight for strategic decisions

All integrated. All working together.

4. Human skills must evolve

The strategic advantage will belong to organizations that have not only invested in the right technology but have also retrained their people and redesigned their processes around a human-machine teaming model.

Your security team needs to:

  • Understand how AI works
  • Interpret AI-generated findings
  • Tune AI models for your environment
  • Make strategic decisions AI can’t
  • Govern AI systems responsibly

The bottom line

AI is now fighting AI in cybersecurity.

The battles happen at machine speed.

Humans can’t keep up.

The choice is simple:

Adopt AI defense—or become a statistic.

Because when AI attacks in seconds and traditional defenses respond in hours…

You’ve already lost.

Category: CyberSecurity - My Journey

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

  • March 30, 2026 by ndiki The Frequency of Desperation
  • March 30, 2026 by ndiki Can AI Fix Security Problems Before Humans Even Notice
  • March 23, 2026 by ndiki The Corner Office
  • March 23, 2026 by ndiki AI Is Now Fighting AI: What This Means for Cybersecurity
  • March 16, 2026 by ndiki The Road to Nowhere
April 2026
M T W T F S S
 12345
6789101112
13141516171819
20212223242526
27282930  
« Mar    
© 2026 porkeynote

Powered by
...
►
Necessary cookies enable essential site features like secure log-ins and consent preference adjustments. They do not store personal data.
None
►
Functional cookies support features like content sharing on social media, collecting feedback, and enabling third-party tools.
None
►
Analytical cookies track visitor interactions, providing insights on metrics like visitor count, bounce rate, and traffic sources.
None
►
Advertisement cookies deliver personalized ads based on your previous visits and analyze the effectiveness of ad campaigns.
None
►
Unclassified cookies are cookies that we are in the process of classifying, together with the providers of individual cookies.
None
Powered by