Artificial Intelligence June 23, 2025

Why legacy cybersecurity breaks under AI-scale attack volumes

The numbers are bad enough already. Kaspersky says it saw 3 billion malware attacks in 2024, along with 467,000 malicious files detected every day. It also found that 45% of passwords can be cracked in under 60 seconds. Even that misses the bigger sh...

Why legacy cybersecurity breaks under AI-scale attack volumes

Traditional cybersecurity is breaking under AI-scale attacks

The numbers are bad enough already. Kaspersky says it saw 3 billion malware attacks in 2024, along with 467,000 malicious files detected every day. It also found that 45% of passwords can be cracked in under 60 seconds.

Even that misses the bigger shift.

Attackers now use AI to compress the whole intrusion chain. Recon, exploit generation, phishing, privilege escalation, persistence, exfiltration. Work that used to take a skilled operator hours or days can now be pushed through a semi-automated loop. Security teams still leaning on signatures, static rules, and perimeter controls are reacting to attacks that change faster than those systems can classify them.

For developers and AI engineers, that changes the job. If you ship LLM features, run ML pipelines, or expose inference APIs, the security model has changed under your feet.

Why the old stack is struggling

A lot of enterprise security still assumes threats are known, or at least slow enough to model with rules. That assumption is breaking.

Signature-based detection works when malware families stay stable. AI-generated malware doesn't need to stay stable. It can rewrite itself, repackage payloads, alter delivery patterns, and generate endless variants that look different at the byte or syntax level while doing the same thing. Static scanners lose that race.

Phishing has the same problem. Attackers can now generate fluent, localized email copy tuned to a target's role and region. The old tells, bad grammar, awkward phrasing, obvious formatting, are fading fast. Add voice cloning and an "approve this access request" call gets a lot more convincing.

Then there's the part many security programs still treat as secondary: AI systems create new attack surfaces of their own.

Prompt injection is the obvious example. If your model can call tools, query internal systems, or act on a user's behalf, model behavior is part of your security boundary. That's a soft boundary. Weight tampering, poisoned fine-tuning data, leaked prompts, unsafe retrieval pipelines, and over-permissioned agents all widen the blast radius.

A lot of teams are bolting AI features onto architectures that were never designed for them. That's showing.

AI attacks move faster because the workflow is automatable

A better way to frame this is workflow compression.

A capable attack chain now looks like this:

  1. Reconnaissance: scrape public docs, fingerprint frameworks, inspect open-source repos, find exposed endpoints, read employee social profiles.
  2. Exploit synthesis: generate or adapt payloads for the target stack, including SQL injection variants, deserialization bugs, or credential theft tooling.
  3. Delivery: produce tailored phishing emails, fake login pages, or support chat messages in the target's language and tone.
  4. Execution and movement: identify weak IAM roles, overbroad service accounts, forgotten internal APIs, then pivot.
  5. Exfiltration: tune traffic patterns to dodge simple anomaly thresholds, sometimes using channels like DNS tunneling or steganography.

AI still doesn't do all of this perfectly on its own. Often it can't. But it reduces the skill and time needed to do enough of it well enough. That's the point. Security failures don't require elegance. One weak path is enough.

Developers are part of the front line now

If you build software, some of this is your problem whether security teams like the org chart or not.

The old split, app team ships features and security scans later, breaks down quickly in AI-heavy systems. By the time a separate review happens, the risky parts are already baked in: unvetted training data, broad model permissions, logs full of sensitive prompts, retrieval systems indexing things they shouldn't, inference endpoints exposed behind weak auth.

A reasonable baseline now includes:

  • Security checks in the ML pipeline, not just the deployment pipeline
  • Data validation at ingestion time to catch poisoning or malformed records
  • Model behavior audits in production, especially for drift toward unsafe or policy-violating output
  • Tighter IAM around AI services, with short-lived credentials and narrow scopes
  • Monitoring at the API layer, because that's where a lot of reconnaissance and abuse shows up first

"Shift left" is tired phrasing, but the underlying point still holds. If you wait until runtime to discover a poisoned model or a prompt-injectable agent with tool access, you're already in trouble.

Shadow AI is boring and expensive

The flashy stories focus on autonomous cyber weapons and nation-state tooling. Fine. One of the most immediate enterprise risks is much less dramatic: shadow AI.

Employees paste credentials, customer data, source code, incident logs, internal design docs, and proprietary prompts into third-party AI tools every day. Sometimes they're trying to move faster. Sometimes they're debugging. Sometimes they don't realize they've crossed a policy line.

That creates a leak path for:

  • PII
  • API keys and tokens
  • internal architecture details
  • training data
  • model weights or prompt logic
  • customer contracts and regulated content

Security teams need visibility into AI service usage, not just endpoint malware. Developers also need to stop assuming that "internal" prompts, notebooks, or support scripts stay internal.

Behavior beats signatures

There's no magic product category here. But one pattern keeps holding up: behavior-based detection is more useful than signature matching.

If attackers can mutate code and text endlessly, defenders need to watch for actions that are rare, mistimed, or inconsistent with how a system normally behaves. That includes:

  • unusual IAM changes
  • unexpected model endpoint invocations
  • inference requests from odd geographies or identities
  • bulk retrieval activity against sensitive vector stores
  • new service-to-service communication paths
  • bursts of failed prompt or tool calls that look like probing

A simple example is watching cloud control-plane activity for rare API calls. If your application never creates IAM users in production and suddenly issues CreateUser or AttachRolePolicy, that matters more than whether the payload hash matches yesterday's malware sample.

import boto3
from collections import deque
from datetime import datetime, timedelta
from time import sleep

client = boto3.client("cloudtrail")
seen = deque(maxlen=1000)

def fetch_events():
end = datetime.utcnow()
start = end - timedelta(minutes=5)
response = client.lookup_events(
StartTime=start,
EndTime=end,
MaxResults=50
)
return response.get("Events", [])

def suspicious(event):
rare_apis = {"CreateUser", "AttachRolePolicy", "InvokeEndpoint"}
return event["EventName"] in rare_apis

while True:
for event in fetch_events():
if suspicious(event) and event["EventId"] not in seen:
seen.append(event["EventId"])
print(f"ALERT: {event['EventName']} by {event.get('Username', 'unknown')}")
sleep(60)

Teams that want broader coverage are also putting canary tokens and honey endpoints around high-value AI services. A decoy endpoint that looks like an internal model API can tell you a lot about reconnaissance activity. If something touches it, you probably want to know right away.

The AI supply chain belongs in security engineering

Most teams are used to dependency scanning for app packages. AI systems add another layer: models, weights, datasets, embeddings, prompt libraries, and containerized inference stacks.

That supply chain is messy. Public model repositories are useful, but they also leave room for tampered weights, poisoned metadata, and quiet provenance failures. Pulling models into production the way you pip install random packages is reckless.

At minimum:

  • verify checksums
  • prefer signed artifacts
  • use provenance tooling such as Sigstore where available
  • track dataset lineage
  • isolate fine-tuning inputs
  • review who can publish or replace models internally

The awkward truth is that a lot of ML platforms still have weaker governance than ordinary application deployment. Fast-moving AI teams often accept that tradeoff to ship. Sometimes they have to. They should at least be honest about the risk.

Zero-trust matters more when models can take actions

AI microservices often end up overprivileged because teams optimize for convenience. The model needs to call search, a database, a ticketing system, maybe a code executor, maybe a CRM. Every connection is another opportunity for abuse.

If your model or agent can perform actions, service identity matters a lot more. Mutual TLS between services, scoped tokens, strict tool permissions, and hard limits around what an agent may do are now basic hygiene. So is separating read-only retrieval from write-capable actions.

A lot of bad agent design comes from one lazy decision: giving the model broad access and hoping prompt instructions keep it under control. They won't.

Defenders will use AI too

Security teams are already using AI for triage, anomaly detection, and automated investigation. That helps, especially when alert volume is absurd. It doesn't remove the harder work of reducing attack surface, cleaning up IAM, validating data, and instrumenting systems well enough to observe abuse.

There's also a trust problem. If a security model flags a request as malicious, teams still need some explainability. "The model thinks it's bad" won't hold up in incident response or compliance review. Behavior-rich signals, audit trails, and plain evidence still matter.

The old security stack isn't dead because AI exists. It's failing because too much of it assumes stable threats, clear boundaries, and slower attackers. Those assumptions are gone.

For engineering teams, the practical response is clear enough: treat models, prompts, training data, and agent permissions as production attack surface. Monitor behavior, not just binaries. Lock down AI services like they can be abused, because they can. And stop shipping LLM features with the same security posture you used for a CRUD app in 2019.

Keep going from here

Useful next reads and implementation paths

If this topic connects to a real workflow, these links give you the service path, a proof point, and related articles worth reading next.

Relevant service
AI automation services

Design AI workflows with review, permissions, logging, and policy controls.

Related proof
Marketplace fraud detection

How risk scoring helped prioritize suspicious marketplace activity.

Related article
The security startups from Startup Battlefield that actually track new attack surfaces

TechCrunch’s Startup Battlefield surfaced a useful cluster of security companies this week, and the pattern is clear. The better ones aren’t slapping AI onto old product categories. They’re built around a simpler fact: models, agents, and synthetic m...

Related article
Congress weighs a 10-year ban on state AI regulation enforcement

Congress is weighing a proposal that would block states and cities from enforcing laws that regulate AI systems for a decade. The language, inserted by Sen. Ted Cruz into a GOP megabill, would tie that preemption to roughly $42 billion in broadband f...

Related article
Automation vs Agentic AI: Different Systems, Different Failure Modes

Yuval Noah Harari’s latest keynote lands on a distinction the industry still blurs when it’s convenient. Automation and agentic AI are different kinds of systems with different failure modes. If teams treat them as the same thing, they ship risk they...