Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
#JaneStreetBets$7BonCoreWeave
What just happened between Jane Street and CoreWeave is not a normal “partnership announcement” — it is a signal of how aggressively AI infrastructure is being absorbed into global capital markets.
A multi-billion dollar exposure package has effectively formed between the two sides, combining long-term compute commitments and an equity investment. This immediately places the deal among the largest crossovers between a trading firm and an AI infrastructure provider in recent memory.
---
What makes this deal unusual
Jane Street is not a typical enterprise customer. It is a high-frequency, quantitative trading powerhouse where infrastructure directly translates into trading edge. Their decision to lock in large-scale AI compute access signals one thing clearly: AI systems are now becoming core trading infrastructure, not experimental tools.
Instead of using compute casually, Jane Street is effectively securing long-term access to GPU-heavy AI systems that can support:
large-scale data modeling
market simulation systems
real-time strategy optimization
machine learning–driven trading research
This pushes AI infrastructure providers into a new category: financial-market-grade compute backbone.
---
Why CoreWeave is at the center of this shift
CoreWeave has grown rapidly because it is not trying to compete as a general-purpose cloud provider. Instead, it focuses on GPU-dense, AI-optimized workloads — exactly what frontier AI systems and quantitative firms require.
As demand for compute continues to rise, providers like CoreWeave are seeing a structural shift from short-term usage to long-term contracted capacity. This means customers are increasingly locking in supply ahead of time rather than renting it on demand.
---
The real signal behind the scale of the deal
This type of arrangement reflects three deeper structural shifts:
1. AI is becoming trading infrastructure
Quant firms are no longer relying only on traditional statistical systems. They are integrating AI models into core decision-making pipelines.
2. Compute is now a strategic asset
Securing GPU access is starting to resemble securing energy or raw materials in traditional industrial cycles.
3. Financial firms are becoming AI-native
Elite trading firms are building internal AI systems at scale instead of treating AI as an external tool.
---
Why markets are paying attention
Investors are watching this trend closely because it signals:
sustained demand growth for AI compute
increasing reliance on long-term contracts
stronger revenue visibility for infrastructure providers
continued pressure on GPU supply chains
AI infrastructure is gradually behaving less like software and more like essential utility capacity.
---
The bigger picture
What is happening here is part of a broader convergence:
AI labs need compute for training models
enterprises need compute for deployment
trading firms need compute for market intelligence
All of them are converging on the same infrastructure layer.
And that layer is becoming one of the most important bottlenecks — and opportunities — in the entire AI economy.
---
Final takeaway
This is not just a large deal between two companies.
It is another confirmation that AI compute has become core financial infrastructure, and that access to it is now a competitive advantage at the highest levels of global markets.
The next phase of AI competition will not only be about better models — it will be about control, allocation, and optimization of compute itself.