Published on

Build AI Scales to 100,000 Hours as Data Scaling Becomes Robotics’ New Frontier

A high-resolution mosaic representing the Egocentric-100K dataset by Build AI, where each individual "pixel" is a distinct first-person video stream. The central view shows a factory worker's gloved hands performing a precise assembly task, framed by the millions of other data points that make up the 100,000-hour collection.
A massive scale-up: Build AI's interactive visualization of Egocentric-100K. Every pixel in this image represents one of the 10.8 billion frames captured by 14,228 factory workers. This data is intended to help robotic models learn complex manipulation tasks by watching human experts.

Just weeks after positioning itself at the forefront of the "physical AI" race, Build AI is accelerating. The startup, led by 18-year-old Columbia dropout Eddy Xu, has released Egocentric-100K, a dataset featuring over 100,000 hours of first-person factory footage. The release is accompanied by news of a significantly expanded funding round, bringing the company’s total raise to $15 million.

The announcement comes at a pivotal moment for the industry. Only a month ago, Build AI made headlines for open-sourcing 10,000 hours of video, then the largest dataset of its kind. By increasing that volume tenfold in such a short window, Build AI is signaling that the era of "data scaling" for robotics has moved from theory to aggressive execution.

The $15 Million "Back to Work"

Founder Eddy Xu revealed the updated funding total on X, quote-tweeting his previous September announcement of a $5 million seed round with a simple correction: "$15m*". The round, which was originally led by Abstract, Pear, and HF0, now includes a high-profile roster of angel investors and industry veterans.

Among those credited in the update are Balaji Srinivasan (author of The Network State and former CTO of Coinbase), Guillermo Rauch (CEO of Vercel), and Thomas Wolf (Co-founder of Hugging Face). The lean nature of the announcement—concluding with Xu’s signature "back to work"—reflects the company’s stated philosophy that speed and research contribution are the primary metrics of success.

100,000 Hours of "Active Manipulation"

The newly released Egocentric-100K dataset represents a massive leap in scale and density. According to the technical details posted on Hugging Face, the dataset includes:

  • Total Duration: 100,405 hours
  • Participants: 14,228 factory workers
  • Total Frames: 10.8 billion
  • Storage Size: 24.79 TB
  • Format: 256p, 30fps H.265 video captured via "Build AI Gen 1" monocular head-mounted devices.

Crucially, Build AI claims the dataset maintains "state-of-the-art" density in terms of hand visibility and active manipulation. Unlike general "in-the-wild" egocentric data, which might include walking or idle observation, this data is captured exclusively in factory settings where workers are performing skilled, economically useful tasks.

A bar chart titled "Egocentric Datasets" comparing the size of various robotics research datasets in hours. The "Egocentric-100K" bar is the longest at 100,000 hours, followed by "Egocentric-10K" at 10,000 hours. Other notable datasets listed include Ego4D at 3,670 hours and EPIC-KITCHENS-100 at 100 hours.
Data Scaling in Action: Build AI's Egocentric-100K dataset is now an order of magnitude larger than its predecessor and dwarfs other established "in-the-wild" robotics datasets like Ego4D and Assembly101.

The "Emergent" Bridge to Robotics

The timing of this release is particularly relevant following a recent breakthrough by Physical Intelligence (Pi). In a recent research update, Pi demonstrated that Vision-Language-Action (VLA) models exhibit an emergent property: as they scale, they spontaneously learn to align human movements with robot actions.

This finding validates Build AI’s core strategy. If a sufficiently large model can "translate" human hands into robot grippers without specialized hardware, then the bottleneck for general-purpose robotics is no longer just robot hardware, but the sheer volume of high-quality human demonstration data.

Xu also hinted at new data collection modalities, sharing footage of workers equipped with wrist-mounted cameras. "Wrist cam changes everything," Xu noted, suggesting that Build AI is moving beyond head-mounted views to capture the close-up, high-fidelity manipulation data that models like Pi’s π0.5\pi_{0.5} thrive on.

A "Delusional Ambition"

The trajectory of Build AI is as much a story of its founder as its data. Xu, who turned down over $25 million in equity to pursue the company, has a track record that spans from winning business world championships to building "chess glasses" used by high-profile creators.

His team, composed of researchers who left top labs and academia, operates under the belief that "the only acceptable speed is as fast as physically possible." With Egocentric-100K, they are testing the limits of that speed, betting that the path to "physical superintelligence" is paved with billions of frames of human labor.

As the industry watches to see if this data leads to a "ChatGPT moment" for humanoids, Build AI has provided the academic and research community with the largest sandbox yet to find out.

Share this article

Stay Ahead in Humanoid Robotics

Get the latest developments, breakthroughs, and insights in humanoid robotics — delivered straight to your inbox.