Futures
Access hundreds of perpetual contracts
TradFi
Gold
One platform for global traditional assets
Options
Hot
Trade European-style vanilla options
Unified Account
Maximize your capital efficiency
Demo Trading
Introduction to Futures Trading
Learn the basics of futures trading
Futures Events
Join events to earn rewards
Demo Trading
Use virtual funds to practice risk-free trading
Launch
CandyDrop
Collect candies to earn airdrops
Launchpool
Quick staking, earn potential new tokens
HODLer Airdrop
Hold GT and get massive airdrops for free
Pre-IPOs
Unlock full access to global stock IPOs
Alpha Points
Trade on-chain assets and earn airdrops
Futures Points
Earn futures points and claim airdrop rewards
A joke about "training AI with two books" just happens to illustrate that computing power is the key.
This joke, turned on its head, punctures the fantasy of “minimal data”
Elon Musk joked that Grok was trained with “just these two books,” “done”—a typical Musk-style quip. What he’s mocking is the fantasy that you can build competitive AI without massive compute. The reality is that xAI is pushing training forward on a massive GPU cluster. Which two books he referred to wasn’t specified (and that part is actually irrelevant), but the meaning is crystal clear: in areas where scaling laws still dominate, he’s making fun of overly simplified narratives.
This tweet sparked polarized reactions. Some people took it as a hint of efficient training; others saw it as more like attention-shifting—what xAI is actually doing is scaling up reinforcement learning on its own Colossus infrastructure. Grok scores (for example, Grok 3 Think achieving 93.3% on AIME) come from compute and training paradigms, not from “reading two paperback books.”
Compute wins; “data minimalism” doesn’t hold up
The spread of this tweet exposes the gap between “catchy viral slogans” (“just two books!”) and “the real lever for building strong models” (massive training on mega-scale clusters). As scrutiny on training data compliance and leakage increases—such as Stanford’s recent documentation of models repeating copyright-protected novels—this becomes even more critical.
xAI is positioning Grok 4 as the strongest level of agentic reasoning by applying RL at the pretraining scale. Unlike OpenAI and Anthropic’s more cautious approach, xAI is joking about “efficiency” while actually delivering multimodal tools. Interpreting this tweet as the popular view of “open source” or an “efficiency revolution” is mostly emotional anticipation—xAI’s $6 billion Series C primarily goes to infrastructure, not “dataset minimalization.”
This also creates a mismatch between pricing and narrative. If the market over-focuses on cost efficiency, it may overlook the higher weight of the compute moat. xAI has a relative advantage in infrastructure; companies like Meta may fall behind on inference depth if they can’t match the same scale of RL and training compute.
Conclusion: The real variable hidden by this joke is xAI’s compute lead. Builders who haven’t shifted toward scalable RL are already behind; investors betting on compute and infrastructure moats are still in an early stage; enterprise buyers adopting Grok’s agentic tools now will be better positioned than rivals still clinging to the “minimal data” myth.
Importance: Medium
Category: Technical insights, industry trends, market impact
Verdgment: The timing to enter this narrative is now: for investors and enterprise buyers betting on compute and RL infrastructure, it’s “early advantage”; for builders still insisting on a data-minimal route, it’s already “too late.” Those who benefit most in practice are participants who control or have access to large-scale GPU clusters and the RL engineering stack: infrastructure builders and mid-to-long-term funds benefit the most, and enterprise buyers willing to deploy the Grok agent toolchain early are also advantaged. For short-term traders, unless there’s a clear compute-supply catalyst, the marginal advantage is limited.