The Rise of Physical AI: What Actually Works and What You Need to Know

Oraginzations raise billions of dollars for software that acts as a universal brain for robots. Investors are betting on the promise, but the gap between lab demos and production reality reveals what actually works today, what remains years away, and why the boring applications matter more than humanoid hype.

09 Feb 2026

In November 2025, a startup called Physical Intelligence raised $600 million at a $5.6 billion valuation – just one year after raising $400 million at a $2.4 billion valuation. Jeff Bezos wrote the check. So did Google’s CapitalG fund. And these are quite unique numbers. 

What makes it stranger is what Physical Intelligence actually sells. They don’t make robots. They don’t manufacture hardware. They’ve built software – a foundation model called π0 (pi-zero) that acts as a kind of universal brain for robots. Give it a camera feed and a command like “fold this shirt,” and it figures out how to move robotic arms to actually do it. Different robot, same brain.

The question everyone’s asking now: Is this the next big thing, or is this 2021-style crypto hype dressed up in a lab coat? At Sphere, we’ve been tracking this development closely – and the answer isn’t straightforward.

What Physical AI Actually Means

Let’s start with what the term means, because “Physical AI” sounds like marketing jargon – and it partly is. Jensen Huang from NVIDIA popularized it at CES 2025, calling it “the next big thing for AI.” The basic idea: take the AI that’s been getting better at understanding language and images, and make it control machines that interact with the physical world.

Traditional robots work through explicit programming. A car assembly robot knows: “When part X arrives at position Y, weld at coordinates Z.” It’s deterministic. Change the part slightly, and the robot fails. You need engineers to reprogram it.

Physical AI is different. These systems use foundation models – the same underlying technology as ChatGPT – trained on both digital data and real-world robot interactions. Instead of programming every scenario, you train the model to understand the general “physics” of manipulation. Show it enough examples of grasping different objects, and it learns what “grasping” means at a conceptual level.

When it works, this is genuinely transformative. One model can generalize across multiple tasks and robot types. That’s what Physical Intelligence demonstrated with π0: the same model controlling seven different robot platforms, performing tasks from folding laundry to assembling cardboard boxes.

The Technical Breakthrough: How π0 Actually Works

Understanding why π0 matters requires understanding what makes it different from previous robot control systems.

The model builds on Google’s PaliGemma, a 3-billion parameter vision-language model that was trained on millions of internet images paired with text descriptions. Physical Intelligence took this base model – which already understands visual concepts and language – and extended it with a technique called “flow matching” to generate continuous streams of motor commands.

Here’s why that’s significant: Previous robot learning systems generated actions one step at a time, like writing one word and then deciding what comes next. Flow matching generates smooth trajectories through action space at 50Hz – 50 complete action predictions per second. This matters because real-world manipulation requires continuous, fluid motion. You can’t fold a shirt with jerky, discrete movements.

The training data came from 10,000+ hours of robot operation across seven different physical platforms – robots with different joints, different grippers, different camera angles. The model saw humans and robots performing 68 distinct tasks in various environments. Critically, it learned not just from successful demonstrations but from corrections, failures, and recovery behaviors.

Physical Intelligence claims this cross-embodiment training is what enables generalization. The model doesn’t learn “move joint 3 to angle 47 degrees” for one specific robot. It learns higher-level concepts: “grasp,” “manipulate,” “orient,” “place.” These concepts transfer across different hardware platforms.

In testing, π0 outperformed baseline models like OpenVLA and Octo on five manipulation tasks. More importantly, after fine-tuning with just 1-20 hours of task-specific data, it could handle complex real-world scenarios like fetching laundry from a dryer, carrying it to a table, and folding each item – recovering when humans deliberately interfered mid-task.

The company open-sourced the model in February 2025, which tells you something about their business strategy: they’re not trying to sell software licenses. They’re proving the technology works to attract partnerships and data-sharing agreements. The real value isn’t in the model weights – it’s in the data collection infrastructure and the next iterations trained on exponentially more diverse robot experiences.

Retrieve Before You Guess.
Verify Before You Act.

RAG + knowledge graphs cut hallucinations and risk.

The Money Flood: Why VCs Are Suddenly Bullish

The funding numbers reveal how dramatically sentiment has shifted. In 2023, robotics funding totaled $6.9 billion. In 2024, it hit $7.5 billion – but concentrated in fewer, larger rounds. The median round size increased while deal count dropped, suggesting investors are getting pickier but writing bigger checks to winners.

Physical Intelligence’s trajectory exemplifies this pattern. From $70 million seed in March 2024, to $400 million Series A in November 2024, to $600 million Series B in November 2025 – each at progressively higher valuations. Total raised: $1.07 billion.

They’re not alone. Skild AI, Figure AI, and similar foundation model startups have collectively raised billions. Even more telling: who’s investing. Jeff Bezos has backed multiple Physical AI companies beyond just Physical Intelligence. Bond Capital, Lux Capital, Thrive Capital – serious growth-stage investors betting that robot control software will be winner-take-most.

Goldman Sachs projects the humanoid robot market alone could reach $38 billion by 2035. Market.us forecasts the broader Physical AI market at $146.8 billion by 2033. These aren’t precise predictions – they’re signals that sophisticated analysts see genuine market development, not vaporware.

But here’s what the projections often miss: the deployment gap. Our analysis at Sphere Inc. suggests the real story lies in understanding where Physical AI actually delivers value today versus where it remains aspirational.

The Reality Check: What Actually Works in Production

Amazon operates over 1 million robots in its fulfillment centers. BMW uses autonomous vehicles within its factories. These are real deployments at scale generating measurable efficiency gains.

But look closer. Amazon’s robots excel at specific, repetitive tasks in highly structured environments. Mobile robots navigate predefined routes. Robotic arms pick items from standardized bins under controlled lighting. These systems work because they operate within carefully engineered constraints.

The gap between these industrial deployments and the vision of general-purpose humanoid robots is enormous – and it’s not closing as fast as funding rounds suggest.

A recent analysis by a16z identified the “physical AI deployment gap”: lab demos achieving 95% success rates drop to 60% in real-world deployment. Not because the technology is bad, but because the real world contains infinite edge cases that no training set captures.

Consider warehouse bin picking – one of the simpler manipulation tasks. Current systems handle structured product categories: packaged goods with consistent shapes, standard lighting, engineered bin presentations. They struggle with arbitrary objects in cluttered environments – the exact scenarios that look impressive in research demonstrations.

Humanoid robots face even steeper challenges. Most deployments remain in pilot phases. They require human supervision for navigation, frequent assistance with dexterity tasks, and constant monitoring for safety. Tesla claims it’s building thousands of Optimus robots for internal use, but independent observers note the demonstrations still show frequent failures that get edited out of promotional videos.

The power problem alone limits practical deployment. Current humanoid robots operate 1.5-2 hours before requiring recharge. Some new designs extend this to 3-4 hours through better motors and lighter materials. But compare that to a human worker’s 8-hour shift. The economics only work if robots can charge during natural downtime or if you’re running multiple shifts with swappable battery packs.

Cost remains prohibitive for most applications. Humanoid robots currently cost $100,000-150,000 per unit. Tesla targets $20,000-30,000 for Optimus, but even that price point requires high volume manufacturing that doesn’t yet exist. The supply chain for specialized actuators, sensors, and components is “non-existent,” according to Elon Musk – a rare moment of candor from typically optimistic promotional statements.

Where It’s Actually Working: The Boring, Profitable Applications

Strip away the humanoid robot hype, and you find genuinely impressive commercial deployments in mundane applications.

Autonomous mobile robots (AMRs) in warehouses have crossed the adoption threshold. Unlike older Automated Guided Vehicles that followed magnetic tracks, modern AMRs use computer vision and AI to navigate dynamically. They avoid obstacles, optimize routes in real-time, and coordinate with other robots without central control.

GXO Logistics deployed these at scale. Walmart automated 100% of regional fulfillment centers using Symbotic’s AI-powered robots. These systems deliver measurable ROI: 25% efficiency improvements, 10% better route optimization, faster order fulfillment.

Collaborative robots (cobots) have found product-market fit in manufacturing. These robots work alongside humans safely, without safety cages. In 2024, global cobot shipments exceeded 50,000 units, up 14% year-over-year. Unlike traditional industrial robots that cost $100,000-500,000 and require extensive setup, cobots start around $20,000-35,000 and can be redeployed to different tasks.

BMW’s German factories use them for tasks requiring precision that humans find tedious: inserting small fasteners, applying consistent torque, holding components in place during assembly. The key insight: these robots don’t replace assembly line workers – they handle the specific subtasks where human dexterity and judgment aren’t required.

Surgical robotics represents another proven application. Intuitive Surgical’s da Vinci system reached 10,488 installed units in 2025, nearly doubling in five years. About 50% of surgeons now perform some robotic surgery, up from 9% in 2012. These systems don’t operate autonomously – they provide superhuman steadiness and precision while the surgeon maintains control.

Autonomous trucking is finally moving from pilots to limited commercial deployment. Waabi plans fully driverless trucks by late 2025. Aurora Innovation and Torc Robotics haul freight for FedEx and Uber Freight on fixed routes in Texas. These aren’t general-purpose vehicles navigating arbitrary roads – they’re optimized for highway driving between specific distribution centers.

The pattern across successful deployments: narrow applications in controlled or semi-controlled environments where the economic value is clear and the edge cases are manageable.

The Three Hard Problems Nobody’s Solving Yet

Despite rapid progress, three fundamental challenges separate impressive demos from reliable deployment:

  1. The Simulation-to-Reality Gap

Current foundation models train largely in simulation. NVIDIA’s Omniverse lets robots practice millions of manipulation tasks in virtual environments before attempting them physically. This accelerates development dramatically – but simulated physics never quite match reality.

A gripper that reliably grasps objects in simulation might slip 20% of the time with real materials. Lighting conditions, surface textures, object deformability – these properties are hard to model accurately. When robots transition to production, performance degrades in ways the training didn’t anticipate.

Researchers are making progress with techniques like domain randomization (training on wildly varied simulated conditions to force robustness) and reality-augmented training (fine-tuning with real-world data). But this remains an active research problem, not a solved engineering challenge.

  1. Long-Horizon Task Planning

Current Physical AI excels at short-horizon tasks: pick this object, place it there, fold this item. These can be accomplished in seconds to minutes with direct sensory feedback.

Multi-step tasks that require planning ahead – “make a sandwich,” “clean this kitchen,” “pack for a trip” – remain extremely difficult. These require hierarchical reasoning: decomposing high-level goals into subtasks, executing each subtask while monitoring progress, handling failures and pivoting to alternative strategies.

Large language models can generate reasonable task plans in text form. But translating those plans into reliable physical execution, with real-time adaptation when things don’t go as expected, is a different problem. The research community is actively working on this through techniques like LLM-based planning combined with foundation model execution, but production-ready solutions remain elusive.

  1. Safety and Reliability

A software bug in a chatbot is annoying. A software bug in a 125-pound humanoid robot moving through a factory is potentially lethal.

Foundation models can hallucinate – generating plausible but incorrect outputs. In language models, this produces fabricated citations. In robot control, it could produce dangerous actions. A manipulation policy that 95% of the time correctly identifies a water bottle but 5% of the time confuses it with a detergent bottle is unacceptable when the robot might hand it to a human to drink.

Traditional robotics relied on formal verification: mathematical proofs that the system won’t violate safety constraints. Modern AI systems are fundamentally probabilistic – they don’t provide hard guarantees. This creates genuine tension between capability and certifiability.

Researchers are exploring hybrid approaches: AI for perception and high-level planning, traditional control theory for safety-critical execution layers. But regulatory frameworks are only beginning to grapple with these questions. The EU’s AI Act classifies many Physical AI systems as high-risk, requiring rigorous assessment. Standards will likely become more stringent, not less, as deployment scales.

Should You Actually Care About This?

Let’s make this practical. If you’re making decisions about whether Physical AI matters to your business, here’s the realistic timeline:

Immediate (2026-2027): Narrow industrial applications If you run warehouses, distribution centers, or manufacturing operations, Physical AI is relevant now. The technology works for structured environments and repetitive tasks. Companies like GXO Logistics, Amazon, and BMW aren’t running pilots – they’re scaling deployments that deliver measurable efficiency gains.

The decision framework: What tasks consume significant labor, involve predictable environments, and have clear success metrics? If you can check those boxes, current-generation AMRs or cobots likely have positive ROI within 2-3 years.

Medium-term (2027-2029): Expanded semi-structured applications Collaborative robots handling more complex assembly tasks. Delivery robots for local logistics. Service robots in hospitality and healthcare settings with human supervision. These applications require more sophisticated perception and manipulation than today’s systems, but they’re grounded in real needs with willing customers.

Watch for three signals: insurance companies offering coverage for autonomous systems at reasonable rates; regulatory agencies approving wider deployment; and most importantly, companies quietly scaling pilots to production without press releases. When deployment becomes routine rather than newsworthy, the technology has crossed the chasm.

Long-term (2030+): General-purpose systems Consumer humanoid robots, fully autonomous vehicles in arbitrary environments, robots performing complex household tasks – these remain aspirational despite aggressive company timelines.

Tesla’s claim of 10 million Optimus units annually by 2027 should be viewed as a goal, not a forecast. Current production is measured in thousands, primarily for internal testing. The supply chain, manufacturing capacity, and regulatory approvals required for millions of units per year don’t exist.

That doesn’t mean it won’t happen – it means it will take longer than promotional videos suggest.

The 5 Pillars of Implementing a Successful AI Strategy

Download our latest e-book to learn how AI and data strategies can drive smarter decisions, higher efficiency, and stronger customer relationships.

Download

The Investment Thesis: Software, Not Hardware

If there’s one clear signal from the funding patterns, it’s this: investors are betting on software, not hardware.

Physical Intelligence doesn’t make robots. Neither does Skild AI or most well-funded Physical AI startups. They’re building foundation models that will run on other companies’ hardware.

The logic parallels the AI wave of 2023-2024: OpenAI captured massive value not by building better chips but by creating models that run on Nvidia’s chips. In Physical AI, the belief is that whoever builds the best “robot brain” will capture disproportionate value regardless of which company manufactures the physical robots.

This might prove correct. Foundation models exhibit network effects: more robot deployment generates more data, which trains better models, which attract more partners, generating more deployment data. If one company’s model becomes the de facto standard, they could extract significant value through licensing or integration fees.

But there’s a counterargument: robot deployment is far more fragmented than LLM deployment. Manufacturing, logistics, healthcare, agriculture – each domain has specific requirements. A foundation model optimized for warehouse picking might not transfer well to surgical assistance. The winner-take-all dynamics of software might not apply.

Hardware integration matters more than in pure software. The companies that deeply understand manufacturing, supply chains, safety certification, and customer deployment will have defensible advantages. This suggests the ultimate winners might be vertically integrated players – companies that control both the AI and the robotics platform.

Amazon might fit this profile better than Physical Intelligence. So might Tesla, despite its erratic execution. Companies with massive internal deployment (generating proprietary training data) plus hardware manufacturing capabilities plus deep customer relationships might compound advantages that pure-software players struggle to match. For more insights on emerging technology trends and market analysis, visit Sphere Inc.

The Sober Conclusion

Physical AI is real. The technical progress over the past two years is genuine, not hype. Foundation models work for robot control in ways that previous approaches didn’t. The economic drivers are compelling: labor shortages, rising wages, decreasing robot costs.

But separating the signal from noise requires distinguishing between research capabilities and deployment reality. Impressive demos ≠ scalable products. Massive funding rounds ≠ mature markets.

What’s actually happening: Physical AI is following a predictable technology adoption curve. Early deployments in controlled industrial settings deliver clear ROI. These will expand steadily to more complex, semi-structured environments over the next 3-5 years. General-purpose humanoid robots remain a longer-term vision requiring breakthroughs in multiple areas.

The hype around humanoid robots specifically is overblown relative to near-term commercial reality. But the broader trend toward AI-powered automation across physical industries is understated. The boring applications – warehouse robots, autonomous tractors, collaborative manufacturing systems – will generate far more economic value in the next decade than humanoid household assistants.

Should you watch this space? Yes, but with specific focus:

If you’re in operations: Monitor AMRs and cobots. The technology is mature enough for selective deployment. Run pilots, measure results, expand what works.

If you’re in tech or finance: Watch the software layer. Who’s building models that generalize across platforms? Who’s collecting the most diverse training data? These companies are positioning for potential platform advantages.

If you’re a consumer: Ignore the humanoid robot timelines. The technology won’t materially impact your life for years. But autonomous delivery, self-driving vehicles, and service robots in public spaces will become noticeably more common by decade’s end.

The rise of Physical AI is real. It’s just not yet the revolution that funding rounds and promotional videos suggest. It’s the beginning of one – which makes it more interesting, not less, if you’re paying attention to the right signals.

For more analysis on emerging technologies and market trends, contact us.

Jaggedness is the price of being early – and the roadmap for what comes next

So what does jaggedness really mean?

It means we’re dealing with systems that are already powerful enough to surprise us, but not yet robust enough to be trusted by default. The market isn’t irrational for hesitating. It’s responding to the true cost of unreliability.

The frontier opportunity isn’t only “who has the biggest model.” It’s “who can make AI reliably useful at scale” – and who can do it while power, regulation, and trust constraints tighten.

In that world, LLM observability is not a DevOps detail. It’s the precondition for scaling. It’s what turns jaggedness from an unavoidable annoyance into an engineering surface you can measure, reduce, and eventually make boring.

And boring – predictably correct, traceable, governed, monitored – is exactly what enterprise value looks like.

Frequently Asked Questions

Physical AI refers to artificial intelligence systems that control robots and machines interacting with the physical world. Unlike traditional robots programmed for specific tasks, Physical AI uses foundation models – similar to ChatGPT – trained on real-world robot interactions to understand general manipulation concepts. These models can generalize across multiple tasks and robot types without requiring explicit programming for each scenario

π0 (pi-zero) is a foundation model built on Google’s PaliGemma vision-language model, extended with flow matching technology to generate continuous motor commands at 50Hz. It was trained on 10,000+ hours of robot operation across seven different platforms performing 68 tasks. The model learns high-level concepts like “grasp” and “manipulate” rather than platform-specific commands, allowing it to control different robot types with the same brain.

Physical AI is commercially viable for narrow industrial applications in controlled environments. Autonomous mobile robots in warehouses, collaborative robots in manufacturing, and specialized systems like surgical robots are already deployed at scale with measurable ROI. However, general-purpose humanoid robots remain in pilot phases, with significant challenges around battery life, cost, safety, and reliability that prevent widespread deployment.

Investors see Physical AI as following a similar trajectory to large language models. Physical Intelligence raised $1.07 billion total, reaching a $5.6 billion valuation in just over a year. The investment thesis focuses on software foundation models that could become the universal “robot brain” across multiple hardware platforms, similar to how OpenAI created value with GPT models. Goldman Sachs projects the humanoid robot market alone could reach $38 billion by 2035.

Three fundamental problems limit Physical AI deployment: (1) the simulation-to-reality gap, where models trained in virtual environments underperform with real materials and conditions; (2) long-horizon task planning, as current systems excel at short tasks but struggle with multi-step sequences requiring hierarchical reasoning; and (3) safety and reliability, since foundation models are probabilistic and don’t provide the hard guarantees traditional robotics relied on.

Despite aggressive timelines from companies like Tesla, consumer humanoid robots remain years away from practical deployment. Current humanoid robots cost $100,000-150,000, operate only 1.5-2 hours before recharging, and require human supervision. Tesla targets $20,000-30,000 for Optimus, but achieving volume manufacturing requires supply chain infrastructure that doesn’t yet exist. Realistic consumer availability is likely post-2030.

Physical AI delivers measurable results in warehouses (Amazon operates 1+ million robots), manufacturing (BMW uses collaborative robots for precision tasks), and logistics (Walmart automated 100% of regional fulfillment centers). Autonomous mobile robots, collaborative robots, and surgical systems like Intuitive Surgical’s da Vinci represent proven applications generating clear ROI in structured or semi-structured environments.

Traditional robots use explicit programming—when part X arrives at position Y, perform action Z. They’re deterministic but inflexible, requiring engineering work for any changes. Physical AI uses foundation models trained on examples to learn general concepts, allowing the same model to handle new situations and different robot types without reprogramming. This enables generalization but introduces probabilistic behavior that complicates safety certification.

Start by instrumenting one high-value workflow end-to-end with consistent request IDs, full traces (prompt → retrieval → tools → output), and a small evaluation suite tied to business KPIs. Then expand coverage once you can reliably detect and explain failures.