Secuvy

Anthropic Leaked Its Own AI Model – Because Even AI Companies Don’t Know What Data They’re Exposing

A company building the world’s most capable AI model left thousands of sensitive internal files in a publicly searchable data store. No sophisticated attacker was involved. No credentials were stolen. The data simply sat in the wrong place, with nothing in place to detect it.

In late March 2026, a security researcher at LayerX Security found nearly 3,000 unpublished Anthropic documents, including a draft blog post about a new model called Claude Mythos, described internally as posing “unprecedented cybersecurity risks,” sitting in an unencrypted, publicly accessible content management system. Days later, a second exposure surfaced: Anthropic had accidentally uploaded the full source code for Claude Code, 500,000 lines across 1,900 files, to a public package repository instead of publishing only the compiled version.

Anthropic attributed both incidents to human error. That explanation is accurate. It’s also exactly the problem.

What Actually Happened: A CMS Misconfiguration, 3,000 Files, and a Model Nobody Was Supposed to Know About

The first exposure came from a configuration error in Anthropic’s CMS. Draft content like blog posts, internal announcements, and unpublished documents became publicly discoverable before anyone at Anthropic realized it.

Roy Paz of LayerX Security and Alexandre Pauwels of the University of Cambridge found the exposed data store. Among the documents was a draft blog post describing Claude Mythos as “by far the most powerful AI model we’ve ever developed,” with dramatically higher performance than Claude Opus 4.6 across coding, reasoning, and cybersecurity benchmarks. Anthropic confirmed the model’s existence to Fortune after being notified.

The second exposure followed days later. Anthropic published the full source code for Claude Code’s agentic harness to NPM,  a public developer repository, instead of uploading only the compiled binary. The complete codebase, containing details about internal architecture and further evidence of an upcoming model tier, was publicly accessible before being removed.

Two major data exposures in one week. Both were caused by human error. Both involve data that should never have left the organization.

The Real Problem Wasn’t the Model – It Was Ungoverned Data in the Wrong Place

Most of the coverage focused on Claude Mythos itself, like how powerful it is, what it means for cybersecurity, and how it compares to competing models.

That framing misses the more important lesson.

No attacker caused this damage. No vulnerability was exploited. The data was simply in the wrong place, accessible to anyone who looked, and no system raised an alert when it got there.

This is one of the most common and costly data exposure patterns in enterprise environments. It doesn’t require a hacker. It requires only a misconfiguration, a missed step in a release process, or a file landing in the wrong storage bucket. The data does the rest.

What Anthropic needed wasn’t better security policies. It needed a system that continuously monitors what data exists, where it lives, who can access it, and whether that state is appropriate, and raises an alert the moment something lands where it shouldn’t be.

Without that layer, human error is always one click away from a public disclosure.

Why ‘Internal’ Doesn’t Mean ‘Safe’: The Classification Gap Most Organizations Ignore

There’s a common assumption in enterprise data management: data created internally is inherently protected. Internal draft documents, unpublished content, source code repositories, staging environments, these get treated as safe by default because they’re not customer-facing.

The Anthropic incident breaks that assumption cleanly.

The exposed documents weren’t sensitive in the way customer PII or financial records are sensitive. They were draft blog posts, internal planning materials, and source code. In most classification frameworks, this content wouldn’t trigger an alert; there’s no credit card number to detect, no Social Security number pattern to match.

But the exposure caused measurable damage. Stock prices for major cybersecurity vendors fell following the Mythos news. Anthropic’s internal model roadmap became public ahead of schedule. The full architecture of a flagship product was made available to competitors and researchers.

This is the classification gap most organizations carry without knowing it. Pattern-based tools catch structured sensitive data, PII, PHI, and financial records. They don’t catch unreleased intellectual property, draft strategic documents, internal product roadmaps, or source code sitting in a misconfigured storage location.

Classifying this kind of data requires understanding context, not just content. A document doesn’t need to contain a credit card number to be something your organization can’t afford to expose. The risk lies in what the document represents, its relationship to an unreleased product, a strategic decision, a competitive advantage, and not in the text patterns it contains.

If Anthropic Can Miss This, What Are Your AI Pipelines Sitting On Right Now?

Anthropic is one of the most technically sophisticated AI companies in the world. They have dedicated security teams, significant resources, and an acute awareness of data risk, the same awareness that led them to describe Claude Mythos as a cybersecurity risk in their own internal materials.

They still had 3,000 documents sitting in a public data store.

For most enterprises, the data environment is significantly more complex. Data estates span hundreds of storage systems, dozens of SaaS platforms, multiple cloud providers, legacy on-premises environments, and a growing set of AI pipelines, RAG systems, and LLM connections, each of which can retrieve, process, and surface data in ways traditional security controls were never designed to handle.

The question for security and data governance teams after this incident isn’t “could Anthropic have done better?” It’s: 

  1. Do you know, right now, what data exists across storage environments connected to your AI systems? 
  2. Do you know if any of it is accessible to someone who shouldn’t have it? 
  3. Would you know within hours if a misconfiguration placed sensitive files in a publicly searchable location?

For most organizations, the honest answer is no. Not because security teams aren’t paying attention, but because the volume, velocity, and variety of enterprise data have outpaced the tools designed to govern it. 

Static scans conducted quarterly can’t keep up with data estates that change daily. Rule-based classification can’t detect risk in documents that don’t match known sensitive data patterns. Manual processes can’t scale to petabyte-level environments where files are created, moved, and modified continuously.

Better Security Policies Won’t Close This Gap. Continuous Data Intelligence Will.

The Anthropic incident will almost certainly produce internal policy updates, access reviews, tighter release processes, and additional sign-off requirements before content reaches staging environments. These are reasonable responses. They’re also insufficient on their own.

Policies govern known processes. They don’t govern the unknown. They don’t catch the draft document that gets saved to the wrong storage bucket. They don’t flag the file included in a release package because someone skipped a step in the build process. They don’t detect when a data store that was private yesterday became publicly accessible today because a configuration changed somewhere upstream.

What catches those things is continuous data intelligence, a system that knows what data exists across all environments, understands what each file contains and what it represents, and monitors that state in real time. When something changes, when a previously restricted file becomes accessible, when a new document appears somewhere, sensitive content shouldn’t exist; the system detects it and responds.

This isn’t a future capability. It’s what self-learning data classification does today. And it’s what separates organizations that discover their own exposures from those that learn about them from a journalist.

Anthropic’s data governance team found out about their exposed files from Fortune magazine.

What Appropriate Data Governance Would Have Changed?

The Anthropic situation points to three specific gaps that continuous data intelligence addresses directly.

Sensitive content detection beyond pattern matching: a system using contextual classification would identify draft product announcements, unreleased model documentation, and internal roadmap content as high-sensitivity, not because they match a regex pattern, but because of what they represent within the organization’s data context.

Real-time visibility into storage accessibility: continuous monitoring of data store configurations detects the moment a storage bucket or CMS changes from private to publicly accessible, and alerts the relevant team before external parties find it.

Data lineage across release pipelines: knowing which files are present in which release packages, and whether any of those files contain material that shouldn’t be in a public distribution, is exactly the kind of intelligence that would have caught Claude Code’s source before it landed on NPM.

None of these controls requires a policy document. They require a system that understands data at the level of content and context, not just location and permissions.

The Broader Point for Enterprise AI Teams

Anthropic’s situation is an extreme, very public version of something that happens quietly inside enterprises each week. A file ends up somewhere it shouldn’t. A storage bucket gets misconfigured. A pipeline pulls data from an unreviewed source. A release package includes files that were never audited.

In most cases, no one finds out, or someone finds out later, during an audit, after a breach, or when a compliance question can’t be answered.

The Anthropic incident was caught by security researchers paying close attention. Most exposures aren’t caught that way.

The lesson for enterprise AI teams isn’t to lock down systems and slow down release processes. It’s that human error is inevitable, and the only reliable defense against it at scale is a system that watches continuously, understands what it’s watching, and acts before the exposure becomes a story.

Data doesn’t stay where you put it. The question is whether you know where it ends up.

Your team should find out about your gaps before anyone else does. Secuvy continuously discovers, classifies, and monitors enterprise data across all environments, so exposures surface internally, not externally. See how it works at secuvy.ai

Related Blogs

April 09, 2026

Most enterprise AI teams are solving the wrong problem first. They’re optimizing storage speed for data that was never safe or ready to use. At...

April 06, 2026

A company building the world’s most capable AI model left thousands of sensitive internal files in a publicly searchable data store. No sophisticated attacker was...

February 28, 2026

“HUMANS, as you know, make MISTAKES.” And that single fact is enough to unravel everything your ChatGPT Enterprise license promised to protect. OpenAI explicitly promises...

February 22, 2026

If you believe ChatGPT Enterprise, Microsoft Copilot, and Claude are secure for enterprise use, consider these uncomfortable facts: ChatGPT has already suffered a bug that...

February 18, 2026

ChatGPT Enterprise prevents OpenAI from training on your data, but it doesn’t stop sensitive data exposure, unauthorized transmission, or regulatory violations. The moment confidential or...

February 14, 2026

“ALERT: SENSITIVE INFORMATION IS LEAKING FROM YOUR SOURCE TO ANOTHER!” Your over-helpful bot would never say that. That’s because AI does exactly what it is...

February 10, 2026

Did you know that Samsung banned ChatGPT & the use of Gen-AI company-wide in 2023? This decision was undertaken as an internal security incident where...

November 15, 2024

Using Data Classification for Effective Compliance When working toward ISO 42001 compliance, data classification is essential, particularly for organizations handling large amounts of data. Following...

November 12, 2024

Laying the Groundwork for ISO 42001 Compliance Starting the journey toward ISO 42001 compliance can seem complex, but with a strategic approach, companies can lay...

November 07, 2024

A Data Subject Access Request (DSAR) is the means by which a consumer can make a written request to enterprises to access any personal data...

November 07, 2024

VRM deals with managing and considering risks commencing from any third-party vendors and suppliers of IT services and products. Vendor risk management programs are involved...

October 30, 2024

With organizations storing years of data in multiple databases, governance of sensitive data is a major cause of concern. Data sprawls are hard to manage...

October 30, 2024

 There has been a phenomenal revolution in digital spaces in the last few years which has completely transformed the way businesses deal with advertising, marketing,...

October 30, 2024

In 2023, the California Privacy Rights Act (CPRA) will supersede the California Consumer Privacy Act (CCPA), bringing with it a number of changes that businesses...

October 09, 2024

For years, tech companies have developed AI systems with minimal oversight. While artificial intelligence itself isn’t inherently harmful, the lack of clarity around how these...

September 25, 2024

Navigating the Shift in AI Compliance Regulations The latest revisions in the Justice Department’s corporate compliance guidelines signal a significant shift for companies that rely...

September 18, 2024

Introduction The threat landscape around data security evolves each year due to factors like a lack of robust security measures, improper data handling, and increasingly...

August 09, 2024

On July 25, 2024, the European Commission released its Second Report on the Application of the General Data Protection Regulation (GDPR), offering an in-depth look...

August 06, 2024

In today’s fast-paced technological landscape, the intersection of AI, data security, and compliance has become a focal point for enterprises aiming to leverage AI’s capabilities...

July 16, 2024

Today Artificial Intelligence (AI) is a part of our day-to-day activities, and knowingly or unknowingly, it impacts our actions and decision-making. With the growing use...

Prepare for Assessments and Get AI-Ready

Gain visibility into sensitive data, reduce exposure, and produce evidence you can trust without months of deployment or manual effort.