Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Launchpad
Be early to the next big token project
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
#ClaudeCode500KCodeLeak
#ClaudeCode500KCodeLeak — The Inside Story of How Anthropic Accidentally Exposed Its Own AI Tool
Date: March 31, 2026
What Happened?
On March 31, 2026, Anthropic unintentionally exposed the internal source code of its AI coding assistant tool, Claude Code, due to a simple but critical packaging mistake.
The company published version 2.1.88 to the NPM registry, but the release included a large source map file (cli.js.map). This file contained the full original TypeScript source code behind the compiled CLI tool.
As a result:
Over 512,000 lines of code
Across nearly 1,900 files
Became publicly accessible
No exploit, no breach, no hacking — just a normal install process.
How the Leak Actually Happened
The issue came down to a basic DevOps oversight:
A build tool automatically generated a source map file
The .npmignore file failed to exclude it
The source map contained full inline source code
It also referenced a publicly accessible storage bucket
This created a direct path: Install → Extract → Reconstruct → Access full codebase
The leak remained live for around 3 hours, which was more than enough time for it to spread globally.
What Was Inside the Code?
This wasn’t just random code — it revealed the full architecture of a modern AI coding system.
System Structure
~1,906 TypeScript files
Modular architecture with commands, tools, services, hooks, and plugins
Around 180+ internal modules
This showed that Claude Code is a complete development system, not just a simple wrapper around an API.
Hidden Internal Commands
Several commands were discovered that were never publicly documented:
/ultraplan → advanced reasoning workflows
/ctx_viz → context visualization
/security-review → automated security checks
/debug-tool-call → internal debugging
/ant-trace → tracing system
/heapdump → memory diagnostics
These reveal internal capabilities far beyond standard user features.
Security Design
One of the most surprising findings was the level of built-in protection:
Multi-layer validation for command execution
Sandbox restrictions for system safety
Strict permission controls for file access
Cross-platform command handling
This confirms that modern AI tools are built with serious safety engineering, not just functionality.
Monitoring and User Behavior Tracking
The code also revealed internal tracking systems:
Performance and usage monitoring
Feature flag systems
Detection of frustration and profanity in user input
This raised important questions about transparency: Users were not fully aware of how their interactions might be analyzed.
Unreleased System: KAIROS
The biggest discovery was an unreleased background agent system called KAIROS, which included:
Always-running background processes
Memory systems ("autoDream")
Notifications and updates
Automatic repository tracking
This suggests a major shift toward: AI agents that operate continuously, not just on command
What Was NOT Leaked
Despite the scale, critical components remained secure:
No model weights
No training data
No user data
No API keys
The intelligence layer stayed protected on servers.
How Fast It Spread
The response from the internet was immediate:
Leak window: ~3 hours
Thousands of downloads within minutes
GitHub repositories gained tens of thousands of stars and forks
Millions of views on shared archives
Once exposed, the code spread too fast to contain.
Anthropic’s Response
Anthropic confirmed the incident briefly, stating that internal code was included in a release but no user data was affected.
There was:
No major legal escalation
No detailed technical breakdown released immediately
The response remained minimal and controlled.
Why This Matters
For the Industry
This leak revealed how advanced AI coding tools are actually built — something rarely visible in a competitive space.
For Developers
It showed:
Real-world AI systems are highly complex
Architecture matters as much as AI models
Tooling is evolving into full ecosystems
For AI Transparency
The discovery of hidden monitoring systems raises important concerns:
What data is tracked?
How is user behavior analyzed?
What should companies disclose?
Key Lesson for Developers
This entire situation came down to one missing configuration rule.
Best practices:
Exclude .map files from production
Always verify packages before publishing
Disable source maps in release builds
Use automated checks in CI/CD
Small mistakes can lead to massive exposure.
Bottom Line
The #ClaudeCode500KCodeLeak was not a cyberattack — it was a simple internal mistake with massive consequences.
Hundreds of thousands of lines of internal code, hidden systems, and future-facing features are now permanently available online.
In an industry built on secrecy, this incident did something rare:
It revealed how modern AI tools are actually designed behind the scenes.
And that makes it a defining moment — not just an accident, but a long-term case study for developers, companies, and the entire AI ecosystem.