Robotics Investments: More Than a Brain
Sidharth Mishra
·
September 26, 2025

Read: “The End of the Elevator Pitch — why innovation stopped being easy to explain.

As we argued in the piece mentioned above, Internet-era innovation was intuitive, narratable, and investable, even for finance professionals without deep technological expertise. A clever website scaled, an app found its market, and value pooled in a few massive cloud platforms. That playbook no longer travels well. Innovation is increasingly embodied: semiconductors with exotic packaging, biotech platforms with long trial arcs, and now robots, machines that need to see, reason, and move. Robotics is a strong innovation theme for the years ahead, but the leap from clicks to kinematics makes everything messier for anyone trying to understand where real economic moats will form.

In this piece, based on our detailed internal report, we will explore a few sub-segments within robotics that could offer compelling investment opportunities. We will remain in the narratable domain, however, and avoid the specific forward-looking corporate analysis that is central to our investment process.

A Short Primer

Start with a humbling fact: a capable humanoid robot contains dozens, often hundreds, of physical parts. Some are brainy, like compute modules and perception sensors. Many are relentlessly mechanical, like actuators, motors, gear trains, and bearings. Only a handful of these components will ever carry a remarkable, defensible business advantage. Most will compete on scale, yield, and cost, just as car seats and fasteners did in the automotive boom. The winners will not simply be the ones with the "smartest" ideas. They will be the ones who master the pieces that are hard to make, harder to copy, and indispensable at volume.

The market backdrop is gigantic. Some experts have claimed that the global humanoid robot market could reach a staggering $3-5 trillion in annual revenue by 2050, with a cumulative stock of over 1 billion units worldwide. This trajectory is driven by a powerful confluence of severe labor shortages and an aging population in the developed world with rising aversion to immigrant labor, which together create a compelling economic case for automation. Low-end Robots priced less than USD10,000 are already being introduced, and even at the highest end, prices could fall for years to come.

From the current near-zero base, it is foolhardy to discuss the likely CAGR for the next decade, although the claimants have pinned high numbers up to 50% per annum. There is no dearth of skeptics who believe that genuinely useful Robots are a long time away. Regulatory friction, safety failures, and energy bottlenecks could slow the curve despite the promises. More importantly, regional divergence will matter more than for Generative AI. U.S. and EU players may face higher compliance costs, while Chinese or Korean suppliers could scale faster with cheaper, if less regulated, systems. This journey from niche technology to a multi-trillion-dollar market will not be driven by a single breakthrough. It will be the result of relentless engineering across a complex supply chain. The story is not just about the brain; it is about the entire body.

Old Robotics ≠ New Robotics

The first key message when we began writing about the innovation era was that 'old AI ≠ New AI,' and this is even more true for Robotics. Our first article was on this topic of confusion created by the term, which had different meanings historically, and we have written more than a couple of dozen articles under its own category since. 

Briefly, we argued it was a travesty of language that the industry kept the same term, AI, for the generative systems that emerged from the transformer architecture. The old meaning, rooted in expert systems and statistical models, created entirely wrong expectations for the new reality of emergent reasoning. A similar confusion now clouds the word "robotics," threatening to anchor our thinking in a past that has little to do with the future being built.

For decades, the term "robot" has been applied to a vast and varied family of machines, from giant, caged arms on an automotive assembly line to the disc-shaped vacuum cleaner in your living room. But they all shared a common ancestor: the pre-programmed automaton. They executed fixed scripts in controlled environments. A robotic arm welded the same spot on a thousand car doors; a warehouse cart followed a magnetic stripe on the floor. They were marvels of mechanical engineering, but they lacked cognition. They followed instructions; they did not understand them.

The arrival of generative AI, which can be considered a form of machine cognition, has fundamentally disrupted this legacy. The defining feature of these new systems is their ability to unlock general functionality. An old robot required explicit reprogramming for every new task or environmental change. In contrast, a robot powered by a generative model can be placed in an entirely new workspace and infer how to achieve a goal. It generalizes from its training data to solve problems it has never encountered, turning a high-level command like "prepare this station for the next shift" into a sequence of novel actions. This is the critical leap from single-purpose automation to adaptable, multi-purpose autonomy.

At GenInnov, we believe "robotics" will therefore remain a difficult and blurry term to define. In the decades ahead, this new layer of general cognition will be integrated into countless devices that we do not currently refer to as robots. A smart forklift that can navigate a chaotic loading dock, a retail shelf that can autonomously rearrange its own stock, or a surgical instrument that can adapt its technique in real time: are they robots or intelligent machines? The distinction will cease to matter, because the defining feature is no longer a specific mechanical form but the presence of this general, adaptive cognitive function along with their machine bodies.

The Shape of Work to Come

The differentiating factor amongst various Robot concepts and products is not in their cognitive capabilities, but in their physical functionalities. This is the case now, and may remain the case for years, if not forever. 

In other words, the early battle lines are being drawn not around the AI models, which are advancing on a common front, but around the physical body. A robot’s value is determined by its mechanical form, the tasks it can perform, and the environment it can safely navigate. This principle is already established in the broader robotics landscape, which features specialized forms like caged industrial arms, collaborative “cobots,” and wheeled mobile robots in warehouses. It is here, in the world of hardware, that the market is fragmenting into distinct and competing philosophies.

The humanoid enters this landscape not as a replacement for these specialists but as a generalist. Its core value proposition is its physical compatibility with human spaces. A machine with two legs and two hands can navigate stairs, walk down narrow aisles, and use tools designed for people. This is not romance; it is capital expenditure math. A robot that fits the existing world saves companies from having to rebuild their facilities for automation. This is also why its physical components, especially dexterous hands, are so critical. They are among the most difficult subsystems to engineer, and mastery of them represents a significant competitive advantage.

This focus on the physical has led to diverging corporate strategies. Some companies, like Tesla and UBtech, are pursuing deep vertical integration, betting that designing and manufacturing their own core components will provide a long-term cost and performance edge. In stark contrast, players like Agility Robotics and Apptronik are leveraging an ecosystem of specialized suppliers, using best-in-class gearboxes from firms like Harmonic Drive and on-device AI compute from NVIDIA. This differentiation extends even to perception, where Tesla's "vision-first" camera-only approach competes with the "multi-modal" strategy of most others, who fuse data from cameras, LiDAR, and other sensors to create a more robust system.

Of course. Here is the revised section with the new title and framing, weaving in the core arguments from the original text with the requested additions.

Hardware Up, Software Down: Chapter 2

This is not to say the cognitive component is unimportant or lacks differentiation. On the contrary, the innovation in robotic "brains" is breathtaking. Rather, the challenge is one of durable advantage. As we have seen in the AI model space over the last two years, extremely rapid innovation and high copyability mean that the economic benefits of a breakthrough can be fleeting. The pioneers of new cognitive skills may not be the ones who ultimately capture the value, as the industry quickly standardizes on the new state-of-the-art.

The core technology driving this revolution is a class of AI called Vision-Language-Action (VLA) models. These systems integrate camera feeds, tactile inputs, and natural-language instructions into physical behavior, enabling a robot to break down a command like "bring me a wrench" into a series of precise movements. The pace of improvement is furious. In headlines, one reads about the progress of platform giants like NVIDIA, with its Isaac GROOT framework, and Google. Just a couple of days ago, Google's DeepMind unit announced its next-generation robotics model, aiming to dramatically improve generalization for complex, multi-step tasks. They may or may not be massively ahead as a result as of today, but they will not be as others will likely be able to introduce similar innovation in weeks and months into their models if they are important.  

This incredible progress, however, also highlights the central challenge: what can be learned fast can be copied fast. Software is fluid. While it pays wonderfully at scale, it is also exposed to competitive learning. Unless good cognitive models are locked behind branded products or physical forms that competitors cannot copy for whatever reason, they risk commoditization. This dynamic does not make the brain unimportant; it makes it necessary but insufficient for a lasting moat. 

It is a paradox, and a critical one. The robot's brain is improving exponentially; this cognitive explosion is what makes this new era possible. In fact, brilliant software is dragging the stubborn physics of hardware into the future. And yet, the durable moats will be built elsewhere. Lasting advantage will be forged not in the fastest code, but in the slow, difficult mastery of the physical world: iconic branding, flawless supply chains, and innovations in atoms, not just bits.

Where Durable Moats May Form

Investment in robotics innovation may be less about the number of parameters in a cognitive model or the data center compute needed for training. The more durable challenge, and therefore the more defensible moat, lies in the physical world: the impeccable coordination of precision components required to convert digital thought into fluid, reliable action. A small fraction of component manufacturers will likely emerge as the most critical and difficult to copy, creating bottlenecks where immense value can be captured.

The most defensible ground is where engineering precision meets stubborn physics. Today, this is most evident in a robot's "muscles and joints." Actuators, paired with their strain wave gearboxes (also known as harmonic drives), are a prime example. These parts, which enable powerful and precise movement with zero backlash, are notoriously difficult to manufacture at scale and can account for over 30% of a robot's total cost. Another area is the development of dexterous hands, a complex mechatronic challenge requiring the seamless integration of numerous joints and sensors into a compact form. While today's leaders in these niches are establishing strong positions, the hardware itself is still evolving. The list of critical suppliers could look very different in a few years as the physical demands of robotics continue to advance.

Humanoids Are a Chapter; Embodied Cognition is the Book

Humanoids get the headlines, but they are just one chapter in a much larger story: the infusion of cognition into every inanimate object. This shift will transform products more profoundly than even the connectivity revolution did. As intelligence spreads from the cloud to the physical world, the very definition of a "robot" will dissolve. Today's forklift, appliance, and medical device are tomorrow's intelligent machines, giving rise to entirely new brands, business models, and service ecosystems.

This brings us to the central point for investors. Unlike the intuitive software and app investments of the past, successfully navigating this new era will demand the same deep study and difficult learning required to understand the semiconductor industry today. The focus must shift from the size of AI models to the physics of their execution. Lasting value will be found by understanding the bottlenecks where thought is translated into action, expanding the value chain from prompts to payloads.

At GenInnov, this is our work. We publish readable maps like this one, but our process is to go deeper, stress-testing narratives against the realities of manufacturing, supply chains, and physics. We hunt for the durable bottlenecks where scarcity lives. In an embodied future, this is where the most defensible moats will be built, and where real value will be created.

Sidharth Mishra is an India-based analyst working through a consultant with GenInnov 

Related Articles on Innovation