Google has revealed Project Suncatcher, a bold initiative to launch AI data centers into orbit—tapping the endless power of the sun to fuel the next generation of artificial intelligence.
In partnership with satellite operator Planet Labs, Google plans to test whether high-performance computing can thrive beyond Earth. The project envisions clusters of small satellites in low-Earth orbit, each equipped with Google’s custom Tensor Processing Units (TPUs) and powered by near-continuous sunlight.
The first two prototype satellites are scheduled for launch by early 2027, marking the initial step toward what Google calls “orbital-scale computing.”
A new frontier for AI infrastructure
Project Suncatcher is designed to tackle one of AI’s biggest problems—energy. Global demand for computing power has surged alongside large-scale model training, putting pressure on terrestrial data centers that already consume vast amounts of electricity and water for cooling.
By moving AI hardware into space, Google hopes to leverage the abundance of solar energy while escaping Earth’s cooling and land constraints. The satellites would operate in a dawn-dusk sun-synchronous orbit, ensuring almost constant sunlight exposure.
“Space offers unique advantages for sustainable, high-density compute,” Google engineers wrote in a blog post announcing the effort. “With uninterrupted solar energy and no need for water cooling, orbital computing could expand what’s possible for AI training.”
Google’s Project Suncatcher: How it would work
Rather than a single giant platform, Suncatcher relies on clusters of small satellites flying in tight formation—forming what Google calls “compute constellations.” The spacecraft would communicate using laser-based optical links, achieving data-transfer speeds in the terabits per second range.
A conceptual design described in Google’s technical paper imagines an 81-satellite cluster spanning roughly one kilometer, forming an orbiting data-center array. Each satellite’s TPU units would process workloads collaboratively, similar to how terrestrial cloud clusters operate today.
Challenges ahead
The project’s ambition comes with significant hurdles. Radiation in orbit can damage memory chips, and dissipating heat in a vacuum is far more complex than on Earth. Maintaining precise satellite formations for optical interlinks will also require continuous station-keeping.
Equally daunting are the economics: Google estimates orbital compute could become viable only when launch costs fall below $200 per kilogram, a threshold that may not be reached until the 2030s.
Still, analysts say the potential payoff justifies the experiment. “If orbital computing works, it could redefine how the cloud scales,” noted one industry observer. “It’s a moonshot—but that’s exactly what Google’s X division is known for.”
The bigger picture
Google’s project Suncatcher underscores how tech giants are reimagining data infrastructure amid the AI boom. While its realization remains years away, the concept pushes the boundaries of both space technology and machine learning.
Whether Suncatcher becomes the next frontier of cloud computing or remains a futuristic experiment, it signals one thing clearly: Google’s ambitions for AI now extend far beyond Earth.
Google has been ramping up its data center construction across the United States, with recent projects including a $4 billion AI data center in West Memphis powered by Entergy Arkansas and a $9 billion expansion of its South Carolina data center infrastructure—but Project Suncatcher marks a completely different approach, taking Google’s computing ambitions beyond Earth.

Project Suncatcher: AI Data Centers in Orbit
Overview
Google initiative to deploy AI computing infrastructure in space using satellite clusters powered by solar energy
Partner: Planet Labs
Timeline: First prototypes launch early 2027
Objective: Orbital-scale computing for AI workloads
Key Drivers
Energy demand: Reduce terrestrial power/water consumption for AI training
Solar abundance: Near-continuous sunlight in sun-synchronous orbit
Scalability: Escape Earth’s cooling and land constraints
Technical Approach
Architecture: Small satellite clusters (“compute constellations”) in tight formation
Hardware: Google Tensor Processing Units (TPUs) per satellite
Connectivity: Terabit-speed laser optical links between satellites
Scale concept: 81-satellite cluster spanning ~1km
Major Challenges
Radiation damage to memory chips
Heat dissipation in vacuum environment
Precision formation flying for optical links
Economics: Requires launch costs below $200/kg (projected 2030s+)
Significance
Experimental approach to sustainable AI infrastructure at cloud scale—outcomes uncertain but represents significant strategic bet on space-based computing.
