07

Physical AI Agents

Humanoids and the Post-Labor Economy

The Robots Are Coming

"This time we are the horse" - RethinkX

A comparison between the development of humanoids and the period when gasoline cars replaced horse transportation. Think about that one.

We're stitching together this perspective note on physical agents that we've been working on for several months this summer. We're doing it because they're coming so fast now. "The robots are coming" is about to become a concept more people will hear about in the years ahead.

Those who believe the internet isn't here to stay or is a flop have become fewer in recent years. So have all those who said "I'll never have a mobile phone." However, there are many who haven't heard of or who believe in self-driving agents, digital agents, and the theme of this perspective note - physical agents.

Physical agents are machines or robots that can perform tasks in the real world, such as carrying packages, assisting in the home, or working in factories. Humanoids are a specialized type that mimics human form and function, making them ideal for interacting with humans and doing productive work in our environments. Physical agents are autonomous or semi-autonomous systems. Humanoids are a subcategory, and strictly speaking, self-driving agents are also a subcategory of physical agents, but we treat them as separate agents.

They Don't Cut Corners

Earlier today I painted a stripe on the cabin. The sun was baking and the refrigerator was tempting, and at the very bottom of the plank that needed painting was a plant of the nettle variety. I looked around and stood up. "Done!" I shouted. It's not dangerous, it's simply human, I thought as I put the lid on the paint bucket. And right here we come to a core characteristic of humanoids and physical agents: "They don't cut corners." A humanoid in a few years would have completed even the last part of the plank. They are humanoids, not humans.

Humanoids consist of hardware that resembles a human body. It's made to move into our environments and do the same tasks we do. And over time better, faster, cheaper, and more precisely. We'll go through how we understand this body with all our limitations. The humanoid also has a brain, or an artificial brain. Silicon and computer vision sewn together into something that sees, evaluates, and acts. The human brain is perhaps the only part of the body that cannot be transplanted or replaced without remaining who you are.

The word "Robot" means slave and work and comes from a play in the 1920s. And that's exactly how we see these three agents (self-driving, digital, and physical) of ours. They will lead us into a world we describe as "post-labour" where the slaves (agents) will produce goods and services that we previously produced with hands, feet, and cognitive processes.

1.0 Historical Perspectives

1.1 The Steam Engine and Humanoids

The transition from human labor to machines can be compared to the transition to the humanoids we see working on the factory floor and taking their first steps into homes. But there are both similarities and differences when we use the energy concept to explain efficiency.

Humanoids that mimic human form and movement represent a further development of machines, but efficiency must be assessed based on their design and use cases. Humanoids increase efficiency compared to human labor in similar ways to machines, by utilizing energy better, working continuously, and performing precise movements. Compared to traditional machines, they sacrifice some efficiency for flexibility, but their human-like form makes them ideal for tasks in our environments designed for humans.

The Development of the Steam Engine: From Static to Mobile

Static Phase (early 1700s to early 1800s):

When James Watt refined the steam engine in the 1700s, mines were pumped empty of water and factories got muscles that never tired. Everything was indeed planted in the same place, and the house stank of coal, but production shot up.

Mobile Phase (mid-1800s onwards):

A few decades later, engineers got the idea to put wheels underneath. The result was the locomotive "Rocket," steamships that crossed oceans, and a world that shrank faster than ever before. Goods and passengers rushed off, cities grew, and global trade became "the new normal."

Chatbot to Humanoid (2010-2020):

In the 2010s, the first chatbots appeared. They answered questions, set reminders, and saved you from eternal waiting in customer service. Everything happened safely in an app.

Mobile Phase (humanoid, ca. 2020-2025 onwards):

Now language models are moving into humanoid robots. They see, hear, and perform physical work. Digital agents go from being "trapped" in our digital gadgets to being among us in the real world. Optimus or Figure can, for example, read the recipe for the cake you're going to make and simultaneously fetch the milk from the refrigerator.

So next time you hear someone say "it's just a chatbot," think about how the steam engine started as a pump and ended as the world's first high-speed transport. Technology doesn't stand still.

1.2 Four Human Qualities Versus Machine Development

Throughout almost the entire industrial journey, we've let machines "borrow" our bodies bit by bit. Now we're approaching a point where they will also borrow our hands, thoughts, maybe even understand the mood in the room. Here's a quick look at how far it has come and where it might end.

Strength

When the steam engine came, muscle power was degraded. Today, robot arms lift car bodies, while exoskeletons make an average warehouse worker into a minor superhero without needing to push down more chicken and rice. Only niche tasks (for example, where it's cramped, dangerous, or legally required with humans present) still require human presence. The rest is already taken by machines, or will be taken soon.

Dexterity

Assembling an old watch or sewing a cut on a patient's heart valve has long been humans' job. Robots were either too stiff, inaccurate, and too slow. Now we have surgical robots that sew finer stitches than the medical team can manage by hand, and factory robots that pick tiny AI chips. They still fall through if the job becomes unpredictable - like tidying a child's room or fishing out keys from the bottom of a bag. But every time the AI gets new updates, they get better and better.

Cognition

Mental arithmetic, chess, and diagnosis tables have already been made cheaper, and in many areas better, by software. But the human brain's strength lies in breadth, not just speed: We can connect completely fresh impressions across contexts. For a machine to reach this level, it must become what we call AGI - that is, be able to learn, form hypotheses, and put things in context on its own. Some think we're already there, but if not, very soon.

Empathy

Machines have learned to sound so caring that a customer service bot can respond "I understand this is frustrating" with the right tone. But real empathy requires more than the right tone. It requires a subjective experience of emotions. No algorithm has that today, and maybe they never will. Nevertheless, "simulated" empathy can be good enough that elderly people feel safe with robot visits. The deep human connection (the one where there's no answer key) currently seems to be one of our human qualities that's difficult to replicate.

2.0 Categorization of Physical Agents by Technology and Use Cases

Physical agents are autonomous robots that operate in physical environments. They perceive surroundings, make decisions, and perform actions. They can be categorized based on technology (basic or advanced) and use cases (specific or general).

Technology Categories

  • Rule-based physical agents: Robots with predefined programs for simple, repetitive tasks (e.g., industrial robot arms in assembly line production)
  • Sensor-based with machine learning: Robots using sensors and machine learning to adapt to dynamic environments (e.g., warehouse robots)
  • Advanced: Robots with deep learning, sensor integration, and decision-making ability in complex environments (e.g., self-driving cars or humanoids)
  • Simulation-supported: Physical agents trained in virtual environments like NVIDIA Isaac Sim to optimize performance before execution

Use Case Categories

  • Industry and production: Robots for automation of production, assembly, or quality control (e.g., in the automotive industry)
  • Logistics and transport: Autonomous vehicles and drones for delivery or warehouse management (e.g., Amazon Scout)
  • Healthcare: Robots for surgery, rehabilitation, or support functions (e.g., Da Vinci surgical robots from Intuitive Surgical)
  • Home: Robots like vacuum cleaners or personal assistants (e.g., 1X's NEO or Tesla Optimus)

Leading Companies in Physical AI Agents

NVIDIA

Has the Jetson platform for edge technology and Isaac Sim for simulating physical agents. Focus on integration of AI, sensors, and GPUs for autonomy in robotics and transport. Leader in simulation tools (Isaac Sim) and edge computing for robots. Used in industry, logistics, and self-driving cars. Tesla's partnership with NVIDIA strengthens their position.

Tesla

Investing in physical agents through self-driving cars (Full Self-Driving) and humanoid robots (Optimus). Focus on deep learning and computer vision. Optimus and self-driving technology, but facing regulatory challenges.

Figure

An American startup (recently raised capital in a Series C funding) that has quickly become one of the most exciting players in the humanoid landscape. Their latest model Figure 02 combines proprietary specialized hardware with language models and computer vision, and is built to operate in human environments like warehouses and factories. What makes Figure particularly interesting is the ambition to create a generalist robot - a physical agent that can not only do one specialized task but learn and perform "everything." With investors like Microsoft and Jeff Bezos behind them, the company has both the capital and vision to become a key player in the agent economy.

1X Technologies

Europe's hope, 1X is vertically integrated: they design everything from drivetrain and sensor network to software, and they train their models on internal data collected through thousands of teleoperated missions. With OpenAI, EQT Ventures, and Samsung among investors, the company secured a Series B round of $100 million in January 2024 to build production capacity. If 1X succeeds, they could become the first European-born player to mass-produce physical agents for our homes. Price tag? Around the cost of an affordable car.

Unitree

Chinese company that has gained global attention with its H1 and G1 series of humanoids. The robots are 178 and 130 cm tall respectively, run up to 3.3 m/s and navigate using 360-degree 3D LiDAR and depth cameras. They develop their own motors with replaceable battery, and one unit can work for several hours before charging, while the price is announced to be under $90,000 for H1 and $16,000 for G1.

3.0 Can Digital and Physical Agents Be Considered a New Species?

In biology, a species is a group of living organisms that can produce fertile offspring while being genetically distinct from other groups. Natural selection, first described by Darwin, favors the traits that best fit the environment, and over time new species are formed. AI systems and robots have no DNA, they don't reproduce, and their mutations are due to code changes rather than genes. In a strict Darwinian sense, they are therefore not a species.

Nevertheless, something similar is happening: weaker algorithms are discarded, stronger versions live on, and humans (or other algorithms) copy, combine, and improve them, much like breeders have shaped livestock through artificial selection.

Transhumanists (Ray Kurzweil being the most famous) believe technology can take over the baton from biology. If artificial systems can one day improve themselves faster than we can keep up, we get a form of "technological species." Tesla's humanoid "Optimus" and today's large language models already show how fast iterations can go when machines train on enormous datasets. If a system achieves so-called general intelligence (AGI) and handles its own further development, we break out of the Darwinian framework and into a new evolutionary form driven by code.

The Verdict

By today's biological standard, AI agents and humanoids are not a species. They lack genes, sex, and natural selection in the classical sense. But if we extend the concept to include self-improving, autonomous technologies, they could become the first "post-biological" species. A "life form" that develops on digital shortcuts instead of through slow mutations. Whether this actually happens depends on two things: how far self-improving algorithms are pushed, and whether they one day develop what we can only describe today as a kind of consciousness. Until then, we should perhaps see them as reflections of our biological intelligence, more than a successor.

4.0 Humanoids as Disruption of Economic Labor

When the steam engine came, it took over muscle power. In our time, software has gradually snatched up routine jobs in front of the PC screen. Humanoids do both at the same time: They lift boxes, repair pipes, and send the report afterwards. One and the same machine can pick goods in a warehouse, talk to the logistics system, troubleshoot a sensor, and order spare parts online. The big difference from classic industrial robots is flexibility. A modern robot can be trained for new tasks with software updates, much like downloading an app to your phone.

Erik Brynjolfsson and Andrew McAfee have shown that automation has so far first eaten the routine jobs. Humanoids expand the menu to home help, construction, service, and parts of the social field. Thomas Piketty's concern that returns on capital grow faster than wages gains new weight. Owners of robot fleets can run off with most of the value creation. Economists who envision a post-work economy therefore bring up arrangements like universal basic income or broad ownership in robots as a possible counterweight.

For many, the job is more than the paycheck. It provides identity, community, and routines. When a machine takes over, people risk being left with time but without a clear purpose. Karl Marx called it alienation. Today we can see the contours of a similar emptiness if work disappears without something else taking its place.

The Optimists' World

Robots produce, humans create. Income from automated value creation is distributed via taxes, funds, or basic income. Education shifts toward creativity, empathy, and entrepreneurship.

The Pessimists' Warning

Capital owners keep the robot income, mass unemployment spreads, and we get digital Luddite riots before new institutions are put in place.

The Transhumanists' Mix

We connect to the machines via Brain-Computer Interface (BCI) and become part of the production apparatus again, just in extended form. Then not only the economy changes, but the very definition of being human.

Physical agents are not just another wave of automation. Humanoids lay claim to the entire spectrum of tasks that have traditionally given people wages and identity. If we manage to update economic models, social safety nets, and our own understanding of meaning quickly enough, the result can be more prosperity and more freedom. If we fail, the gap between those who own the robots and those who don't can become so large that the social fabric is torn to pieces. The choice is not between robots or not, but between letting the upheaval control us or controlling it ourselves.

5.0 How Humanoids and Industrial Robots Can Transform the Future

Julius Caesar is known for several quotes, one of them being: "I came, I saw, I conquered." We humans also follow a simple three-step process. We humans see, evaluate, and act. This process lets us do everything from simple things like grabbing a cup and drinking coffee, to complex tasks like playing piano. Humanoids are designed to mimic humans, so they use technology to recreate the same process: see, evaluate, and act.

Humanoids Use Cameras as "Digital Eyes"

Instead of human eyes, they use cameras, radar, and other sensors to capture visual data, such as images or video. Computer vision is a field within AI that lets machines interpret and understand visual data, such as object recognition, environmental mapping, and navigation to capture images of the environment, just like human eyes. For example, Tesla's Optimus can use 8 cameras to see and understand the surroundings.

AI and Inference is "The Brain"

AI is a set of algorithms that learn from large datasets to process visual data captured by computer vision. Inference is the process where AI uses learned patterns to draw conclusions or make decisions. For example, if computer vision sees a cup, AI uses inference to determine that the cup is something that can be gripped, and how it should be gripped. Inference lets humanoids make decisions in real-time.

The Robot Body

The robot body is controlled by motors, actuators, and sensors. When AI has determined what should be done, it sends signals to perform the action. These parts are designed to be flexible and precise, and can have sensors to provide feedback if something requires sensitivity.

Together, these systems let humanoids interact with the world, such as navigating or performing tasks. However, it's still uncertain how well humanoids can match humans' intuitive abilities. This is especially true in complex, varied tasks.

6.0 Voice Commands Are Central for Humanoids

Voice commands have the potential to disrupt finger-based interfaces. The global market for humanoids (expected $38 billion in 2032) and VUI, i.e., speech/voice commands (USD 30.46 billion in 2025) shows strong growth with over 20% CAGR, due to developments in NLP (Natural Language Processing), edge computing, and multimodal AI.

Humanoid robots are built to work side by side with humans in homes, stores, hospitals, and factories. The interaction must therefore feel natural. Pulling up an app or waving a remote control breaks the illusion that the robot is a helpful colleague. Our voice, on the other hand, is the most intuitive interface we have.

When you say "Optimus, lift the package into the van," the robot interprets your intention, finds the package, and performs the task. You avoid having to describe the exact route, location, or grip, just as you would do with another human.

The Technology That Makes Communication Possible

  • Speech recognition: Modern systems pick up words even in noisy warehouse halls and understand dialects far better than a few years ago.
  • Natural language understanding (NLP): Large language models - think Grok or the GPT family - break down the sentence into actions the robot can perform.
  • Context and sensors: Cameras, lidar, and microphones help the robot understand whether "water" means the glass on the kitchen counter or the plant in the living room.
  • Edge computing: Much of the analysis happens locally in the robot, so the answer comes in an instant and private conversations don't have to be sent out to the cloud.
  • Multimodal AI: The robot combines words with gestures and gaze direction. Point at the shelf and say "over there," and it follows your finger.

7.0 A Complete Overview of the Body and Technology

A humanoid robot's body is built from components that mimic human anatomy. Basically, they are based on advanced technology for movement, perception, and control. The most important components are control systems, actuators, and sensors, while the most challenging single area is the hands.

The Head

Contains sensors like cameras for visual perception, microphones for hearing, and speakers for communication. The head may also have lights or other indicators for interaction. For example, Optimus uses 8 cameras to understand the surroundings.

The Torso

Contains the core of data processing, batteries, and other internal components. It functions as a central part for connecting arms, legs, and head, and must be robust enough to support the robot's weight and movement. It's often made of lightweight materials like aluminum or carbon fiber to reduce weight.

Arms and Hands

Essential for manipulation and interaction with the environment. The arms have joints like shoulders, elbows, and wrists, while the hands are particularly complex, with many degrees of freedom in a small area to mimic human finger precision. The most challenging component is hands. Mimicking human hand dexterity is extremely complex. The human hand has 27 degrees of freedom, and creating robot hands that can handle everything from small, light objects to heavy items is a major challenge.

Legs and Feet

The robot's propulsion system. Each leg has joints that mimic hip, knee, and ankle, and in the soles are pressure sensors that let the machine "feel" the surface and maintain balance. Boston Dynamics' "Atlas" humanoid shows how far the technology has come. It jumps up stairs, crawls, and runs over uneven terrain as if it were flat tartan track.

Key Components

  • Actuators: Function as "muscles" and are often electric or hydraulic. This is what provides movement to joints and limbs. Electric actuators are common due to their compact size, while hydraulic ones provide higher force but were more commonly used earlier.
  • Sensors: Include cameras, tactile sensors for touch, gyroscopes for balance, and others to understand the environment. All these sensors provide data to the control systems.
  • Control Systems: Advanced AI and machine learning algorithms that process sensor information and control actions, often based on processors that can handle real-time data.
  • Structural Components: The "skeleton" of the robot, often made of lightweight materials like aluminum or carbon fiber to maintain strength without weighing too much.

8.0 "The Hand is Where the Mind Meets the World"

The human hand is a marvel of engineering with 27 degrees of freedom, allowing for everything from delicate precision work to powerful gripping. For humanoids, replicating this capability is one of the greatest challenges in robotics.

Current Robot Hands

  • Shadow Hand: 24 DoF, can play piano, tie shoelaces
  • Optimus: 22 DoF, still developing piano-playing capabilities
  • Figure 02: 16 DoF, can pour liquid, turn screws
  • Unitree G1: 10 DoF, can grip boxes, press buttons

Challenges

  • Each additional joint requires a motor and sensor, adding weight, cost, power consumption, and wear
  • Balancing complexity with reliability
  • Achieving human-level sensitivity and precision
  • Managing heat dissipation in compact spaces

Despite these challenges, all components are becoming faster, cheaper, and more efficient. And quickly.

9.0 Training and Learning

Physical agents learn in virtual worlds first. In tools such as Isaac Sim, Unity Robotics, and Gazebo, the robot gets to practice thousands of variants of the same task. Everything from maintaining balance on a downhill slope to assembling a cabinet. This way, companies save both time, money, and wear. One day of simulation in the cloud can equal several years of "physical" practice.

After simulation, reinforcement learning (also called "trial-and-error learning") takes over. The algorithms reward each successful movement and "punish" errors. Optimus finds the most energy-efficient gait, Unitree H1 learns to jump higher without losing balance, and Figure fine-tunes finger movements needed to pour coffee without spilling. This phase is often driven on large cloud clusters because the updates require enormous computing power.

9.1 Case: IKEA Billy Bookshelf

Assembling a Billy bookshelf isn't the most advanced thing a human can do, even though it can be frustrating. But the task encompasses almost everything a versatile robot must master: see, sort, grip, screw, hammer, and solve when a screw rolls under the sofa. That's why researchers have begun using the Billy bookshelf as a test of whether a general humanoid is "ready."

Imagine Figure gets the project today. Before it touches a single wooden board, the development team builds a digital copy of the living room in Isaac Sim. Shelf, tools, screws, and manual are scanned in, even the light from the kitchen window is modeled so the cameras see realistic reflections. This is where the robot begins to practice. First recognizing screws and plugs, then lifting the boards without scratching them.

When the error rate is down to human level, about 95% success rate and under one hour per shelf, everything is packaged into the ROS module and sent OTA to the entire Figure fleet. The robot is ready for missions in your home.

9.2 Learning via YouTube?

We used to think YouTube was primarily for cats playing piano or us nerds discussing semiconductors. Now the same video base is becoming evening school for humanoids. A robot like Figure or Optimus can actually learn to assemble a Billy bookshelf just by watching a few quick clips on YouTube.

Here's how it works: The camera in the robot skull and a VLM (Vision-Language Model) divide the video frame by frame, find the hands, screws, and boards, and link each movement to an appropriate joint pattern in the ROS system. The language engine picks up verbal hints like "tighten the screw one more notch" and places them on the timeline. The entire recipe is test-run in a simulator like Isaac Sim, safely, cheaply, and without danger of scratches on the living room floor, before being sent as an OTA update to all robots in the fleet.

The consequence is that when a robot can learn an entire assembly job in an afternoon of YouTube and a bit of simulation, each new task becomes in practice just an update. Today it's the Billy bookshelf. Tomorrow it could just as well be changing brake pads or wallpapering walls.

10.0 Strategy and Categorization

Amara's law says that we often overestimate technology's short-term impact and underestimate its long-term potential. In an investment context, this is another way of saying we need to think long-term. It's also a nice entry point for analyzing technologies' different phases.

Gartner's Hype Cycle Applied to Humanoids

  1. Technology Trigger (1990s-early 2000s): Early prototypes like Honda's P2 and P3 in 1996-1997, ASIMO in 2000
  2. Peak of Inflated Expectations (mid-2000s-early 2010s): SoftBank's Pepper in 2014, media and investors embraced visions of robots as integrated part of daily life
  3. Trough of Disillusionment (late 2010s): Challenges like unstable gait, limited battery life, and high production costs became clear. Honda ended ASIMO project in 2018
  4. Slope of Enlightenment (2020-2025): Improvements in AI, machine learning, material science, and sensors have increased humanoids' capabilities. Boston Dynamics' Atlas shows advanced mobility
  5. Plateau of Productivity (2025+): We believe this fourth phase is now entering its final hours. Figure, 1X, and Tesla Optimus have shown exponential progress on hardware and software

10.1 Musk's Quote on Humanoids

The quote from the TED interview in 2022 provides a fascinating starting point: Musk sees Optimus as a natural extension of Tesla's work with self-driving cars, where real-world AI is the key, and he emphasizes that humanoid robots can be "bigger than the car."

Elon Musk likes to call Tesla's cars "robots on four wheels." With Optimus, he takes the statement literally - he wants to make the car bipedal. To succeed, he points to two completely fundamental obstacles: the brain (real-world AI) and the body (large-scale production). If Tesla manages to crack both, robots can become as commonplace as vacuum cleaners.

From Self-Driving to Housework

Tesla has spent ten years getting a car to understand intersections and cyclists. Now the same computer vision must tackle an apartment full of Lego blocks, pets, and people coming home with shopping bags. The road is shorter than you might think.

Scale

Boston Dynamics can build one fantastic robot. The problem is their factory looks more like a film laboratory than a car factory. Tesla can spit out one Model Y every 45 seconds. The same assembly line logic will push the robot price down from well over $100,000 to something an ordinary business, and eventually a household, can easily afford.

10.2-10.4 Global Humanoid Landscape

10.2 American Humanoids

The majority of American companies like Agility Robotics, Boston Dynamics, Figure AI, and Apptronik focus on industrial applications. This reflects the USA's strong position in production and logistics, where there's an increasing need for automation to meet labor shortages and increase efficiency.

  • Agility Robotics: Develops Digit, designed for industrial tasks like warehouse handling and logistics
  • Boston Dynamics: Known for Atlas, designed for industrial and research purposes with ability to perform complex movements
  • Figure AI: Develops Figure 01 and 02 for industrial applications like production, logistics, and retail
  • Tesla: The only company developing Optimus with clear focus on both industrial and service-oriented applications

10.3 European Humanoids

USA got Google, Tesla, Facebook, and Palantir, while Europe got cork screws and GDPR. Hard facts. Can humanoids bring the EU back to a stage of relevant companies shaping the future?

  • 1X Technologies (Norway): Develops NEO Beta for home use, designed to assist in daily tasks
  • NEURA Robotics (Germany): Develops 4NE-1, a humanoid robot helping humans in everyday life
  • PAL Robotics (Spain): Pioneer with TALOS, a bipedal humanoid for research and industrial applications
  • Enchanted Tools (France): Develops Miroka, social robots for health, hotels, and retail

10.4 Chinese Humanoids

No country is building bipedal robots faster than China. According to an industry survey in autumn 2024, there are over 60 Chinese companies that have showcased their own humanoids. The government has decided to give them momentum with the "Robotics + Application Action Plan (2024-27)."

  • Unitree Robotics: H1 for electric car factories, G1 for inspections and monitoring
  • UBTech Robotics: Walker S2 for sorting, quality control, and assembly with 500+ orders from car manufacturers
  • Xiaomi: CyberOne designed for customer service and smart home control
  • Beijing promises to: Double robot density in industry to 500 units per 10,000 employees by 2025, cover up to 30% of R&D costs for companies building advanced automation

11.0 Timeline

Have we succeeded in our telling of perspectives, there now exists an imagination that humanoids are moving rapidly from cautious pre-programmed steps in Boston Dynamics' laboratory to computer-vision humanoids with AI brains that want to contribute to productive work.

Short-term (2022-2027): Prototypes and Early Commercial Use

In this period, humanoids will primarily be used in controlled environments like factories, warehouses, and research laboratories. Focus is on improving AI-driven capabilities like movement, perception, and interaction, as well as starting commercial rollout in niche applications like industry and logistics.

  • Tesla plans to produce around 1,000 Optimus robots in 2025 for internal use
  • Agility Robotics' RoboFab factory has capacity for 10,000 units per year
  • Global estimate for 2025: 3,000-10,000 humanoids, primarily in industrial or research environments

Medium-term (2027-2032): Broader Adoption Across Sectors

Humanoids will expand to sectors like healthcare, retail, and service, where they handle complex tasks like assisting elderly, running customer service, or performing maintenance. Advances in AI and reduced production costs will make them more accessible to businesses and institutions.

  • Tesla aims to scale to 10,000 units in 2026, potentially reaching 10,000 units per month
  • Morgan Stanley estimates 40,000 humanoids in USA alone by 2030
  • Chinese companies targeting rapid scaling with government support

Long-term (2032-2040): Integration into Daily Life

Humanoids will become an integrated part of society, with widespread use in homes, schools, and public spaces. They will be able to perform a broad spectrum of tasks, from personal assistance to collaboration with humans in complex projects.

  • Elon Musk predicts 1 billion humanoids globally by 2040
  • UBS estimates 300 million humanoids globally by 2050
  • Post-labor economy where goods and services can be produced by physical and digital agents

12.0 The Interplay Between Digital and Physical Agents

Until now, most of us have been used to artificial intelligence in the form of a chatbot that answers questions, or an algorithm that recommends the next series. Now this software is getting arms and legs. But the software must talk to the arms and legs. Digital agents: software on our digital gadgets (e.g., planning routes or interpreting images) collaborate with physical agents: robots, drones, or self-driving vehicles.

From Rule-Controlled Remote Control to Digital Symbiosis

  1. Remote Control Phase: A simple web service pumps ready-made commands down to a robot on the floor. Effective, but zero flexibility.
  2. Machine Learning Phase: Digital agents interpret streams of sensor data and send dynamic instructions back. The warehouse robot learns where pallets tend to appear and captures increasingly shorter routes.
  3. Simulation Phase: Platforms like NVIDIA Omniverse and Isaac Sim let the software run millions of trial runs in a virtual factory before a single nut is screwed in reality.
  4. The Seamless Ecosphere: In a smart city, a sensor in the asphalt beeps: there's a hole in the road. A digital agent sends the order, a robot constructor responds, the drone delivers the asphalt bag - and traffic never passes through a human inbox.

Industry Winners

  • Highly efficient production
  • Advanced logistics
  • Surgery and elderly care
  • New jobs in data management, robot training, and system integration

Industry Losers

  • Repetitive warehouse and transport work
  • Simple assembly lines
  • Office jobs that just move data without making decisions
  • Traditional transport and production industries not investing in transformation

13.0 Ethics

When investment theme is humanoids, questions about fear sometimes come up. Fear that a variant from the Terminator films will kill us, and other times questions about whether we'll be scared. Often we answer that the calculator is much smarter than us, but we're not afraid of the calculator. The topic that most often comes up in such contexts is about humanoids and ethical action spaces.

Ethical dilemmas are situations where a person or humanoid faces a choice between two or more actions. No matter what they choose, the choice will violate one or more ethical principles or values. There is no obviously right choice and solution.

Isaac Asimov's Three Laws of Robotics

  1. A robot may not injure a human being or, through inaction, allow a human being to come to harm.
  2. A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.
  3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.

Humanoids will undeniably end up in situations where they must make decisions where the right choice is not unambiguous. Human choices in such situations are often explained with concepts like intuition, experience, or human judgment. For humanoids, dilemmas arise at the intersection between programmed rules, human expectations, and possible developed autonomy in the humanoid itself.

A classic dilemma is situations where the humanoid must choose whom to save. Let's imagine an emergency situation where two people are in danger. However, only one can be saved. How should agents decide who gets priority? Should it base itself on age, health status, or other criteria? This is reminiscent of the famous trolley problem.

13.1 Can Humanoids Exert Group Pressure on Each Other?

Is it possible to imagine situations where humanoids exert group pressure, especially if they are equipped with advanced AI that gives them local freedom in perception and action? Group pressure occurs when individuals (or in this case robots) are influenced by the group's norms, goals, or behavior to change their own actions, even if it conflicts with their individual "logic" or programming.

For this to happen, robots must have:

  1. Social AI: Ability to perceive and react to other robots' behavior, like a form of "digital empathy" or coordination logic
  2. Shared goals or protocols: Groups of robots must share a common framework that creates a "group identity"
  3. Local autonomy: Ability to deviate from central directives so they can adapt to local conditions
  4. Learning ability: AI that adapts based on observations of other robots

Yes, humanoids can exert group pressure in the future, especially with local freedom in perception and action. Scenarios like efficiency pressure in factories, social adaptation in homes, or cultural adaptation in public spaces show how this can happen. This raises fascinating questions about AI ethics, societal impact, and investment opportunities.

14.0 Use Cases

14.1 Nursing

NAV (Norwegian Labour and Welfare Administration) believes Norway will lack around 30,000 nurses by 2035. More elderly, tighter budgets, and fewer working hands point to an extreme need for relief, and this is where the bipedal robots from Tesla, 1X, and Figure come in.

In 2025, these machines are still "prototypes": they can roll a cart, fetch bandages, and show pulse on a screen blink, but small talk isn't something they're quite comfortable with yet. By 2030, the same robot can take the heaviest lifts, transport medicines between departments, turn patients to avoid bedsores, wash rooms without complaining about the cleaning smell, and immediately notify if a resident falls.

Looking to 2035, the first truly flexible "robot nursing assistants" are in place. They distribute morning medicine by scanning barcodes and double-checking dosage against the journal, lead a simple physio session and count repetitions without cheating, open the video link to the doctor while taking blood pressure, and happily chat with lonely residents driven by language models that know the difference between polite small talk and signs of anxiety.

14.2 Home Help

Housework has never been at the top of the wish list, neither for students with piles of dishes nor tired home services that have to make new rounds in record time. That's why many robot companies are sharpening their focus on the living room, kitchen, and bathroom. If they can do that, they can do almost anything.

Management at Tesla, NVIDIA, Figure, and 1X increasingly talk about "labor in abundance." They envision robots practically changing the status of labor from a scarcity factor to something that exists by the meter as far as network cables reach.

Bernt Børnich at Norwegian 1X likes to say that a robot should grow up in a completely normal home, not on a ruler-straight factory. The logic is simple: a living room offers a hundred times more surprises than an assembly line does - loose cables, children's toys, pots boiling over, and a cat that's always in the way. That's the data diversity needed if the robot is to become truly versatile.

14.3 How Could a Bunad App Look in 2027?

Imagine it's May 16, 2027. You fetch your Optimus robot, open "Bunad" in Tesla's new app store, and by the end of the day you have a brand new Hardanger bunad that fits perfectly. Sewn, embroidered, and pressed by a robot that can simultaneously explain why exactly this rose fabric and these colors became popular in the fjord villages 150 years ago.

The app is built to unite traditional craftsmanship with robot precision. The first time you start it, you download the pattern package for your region, say Hardanger or Gudbrandsdalen, and Optimus immediately sets out on the task. With cameras, LiDAR scanners, and precise hydraulic fingers, it takes accurate measurements of you, draws a three-dimensional model, and adjusts the cut down to the millimeter.

Unlike cheap copies ordered from factories far away, the "bunad" app focuses on locally sourced materials and authentic patterns. Through the app, you can order wool from Norwegian sheep, yarn from Selbu, or specially dyed silk thread from local suppliers, and everything is digitally tracked so you always know where the materials come from. The result is a bunad that is both more sustainable and more in line with the craft tradition than mass-produced alternatives.

15.0 Disruptive Investment Strategies and Physical Agents

Our investment universe was defined at the entrance to what we called the disruptive decade (2020-2030). We have analyzed, discussed, and invested in satellites, IoT, drones, robots, artificial intelligence, edge technology, metaverse, eVTOL, self-driving cars, drones, fintech, and energy transformations over the past five years. Often to the laughter of investors and advisors who are more concerned with looking in the rearview mirror than looking forward. We live well with that.

The next five years, high beta will become low beta, and low beta will become high beta, we often say. Large index-heavy and well-established business models will be disrupted by new technology and smaller players. Fortunes will probably change hands in the next five years too.

Our Five Investment Categories

  1. Connectivity: IoT, E-commerce, Satellite, Cloud
  2. Urban Mobility: Electric cars, Self-driving cars, Infrastructure
  3. Machine Revolution: Robotics, Artificial Intelligence, Big Data
  4. Demographics: Generation Z, Healthy living, Aging (BioTech)
  5. Green Deal: Renewable Energy, Circular, Water, Electrification

Physical agents directly and indirectly touch several categories and subgroups. The main category is what we call (3) Machine Revolution. (2) Urban mobility we define as the self-driving agents. However, physical agents require a large ecosystem of suppliers where both (1) Connectivity and (5) Green deal become important. Electricity, batteries, access technology (satellites/radio), sensors, and semiconductors are all enablers for physical agents.

The S-Curve and the Reverse S-Curve

A main theme in this and other perspective notes is that we see an S-Curve development among the three agents. S-curves have a buddy - the reverse S-curve - and it shows how the "loser" is replaced. In this context, the loser is humans' hands, feet, and cognitive processes.

The disruption of low beta companies and humans' contribution to economic growth is underway. Those who succeed in using Industry 4.0 to deliver more for less, faster than competitors, and often in markets that didn't exist yesterday. Those who wait may discover that disruption is fast and brutal, and that the train they should have been on has already left the platform.

Key Takeaways

"This time we are the horse": The comparison to how automobiles replaced horses is apt - but this time it's human labor being replaced by tireless, precise machines.

They don't cut corners: Unlike humans who might skip the last bit of paint or rush through quality control, humanoids will complete every task to specification, every time.

Three-step process: Like humans, humanoids see (computer vision), evaluate (AI inference), and act (actuators) - but with perfect memory and no fatigue.

The hand challenge: Replicating human hand dexterity with 27 degrees of freedom remains one of the greatest engineering challenges.

Virtual training first: Humanoids learn in simulated worlds before touching the real world - one day of cloud simulation equals years of physical practice.

Voice is the interface: Voice commands will disrupt buttons and screens, making human-robot interaction as natural as human conversation.

Global race is on: USA leads in innovation, China in scale, Europe struggles but has strong players like 1X and NEURA.

Timeline to disruption: 2025-2027 for industrial pilots, 2027-2032 for service sectors, 2032-2040 for home integration.

Post-labor economy: When machines handle physical and cognitive work, human economic value must be redefined.

Investment implications: High beta becomes low beta, low beta becomes high beta - fortunes will change hands as disruption accelerates.

"The robots are coming. Not as conquerors, but as slaves - digital and physical agents that will produce the goods and services we once created with our hands, feet, and minds. The question isn't whether they're coming, but whether we're ready for the world they'll create. A world where human labor becomes optional, where creativity becomes currency, and where the biggest risk isn't building them - it's not building them fast enough."

This perspective has been translated from Norwegian to English

Download Original (Norwegian)