{"id":33512,"date":"2026-02-16T05:57:23","date_gmt":"2026-02-16T03:57:23","guid":{"rendered":"https:\/\/qviro.com\/blog\/?p=33512"},"modified":"2026-02-16T05:57:23","modified_gmt":"2026-02-16T03:57:23","slug":"why-most-humanoid-robots-lack-ai-stacks","status":"publish","type":"post","link":"https:\/\/qviro.com\/blog\/why-most-humanoid-robots-lack-ai-stacks\/","title":{"rendered":"Why Most Humanoid Robots Lack AI Stacks"},"content":{"rendered":"<p><span style=\"font-weight: 400;\">Humanoid robots are robots built to look and move like humans. They usually have two arms, two legs, and a head. The goal is simple: let the robot work in spaces designed for people, using the same tools and workflows you already have. Today, many of these robots are marketed as <\/span><i><span style=\"font-weight: 400;\">AI-powered<\/span><\/i><span style=\"font-weight: 400;\">. But when you look closer, most of them do not run a full AI stack. So why is that? In this article, you will learn what an AI stack really means for humanoid robots, and why power limits, safety, data, and control complexity still push engineers toward more traditional solutions.<\/span><\/p>\n<h2><span data-text-color=\"secondary\"><b>What an AI Stack Means for Robotics:<\/b><\/span><\/h2>\n<p><span style=\"font-weight: 400;\">An AI stack combines layered technologies that enable a robot to move beyond fixed routines. At the base, it includes sensors that detect the environment. Above that are systems that interpret what those sensors see. Next are decision-making and planning layers that decide what to do. At the top is adaptive control that adjusts actions in real time based on feedback. In robotics this means perception, reasoning, planning, and adaptive control all work together to enable truly intelligent machines. An AI stack turns raw data into decisions that change how the robot behaves in new situations.<sup>[1]<\/sup><\/span><\/p>\n<h3><span data-text-color=\"success\"><b>Perception and Reasoning:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Perception includes computer vision, depth sensing, and audio processing that help a robot understand its surroundings. For real autonomy these systems must not just sense but <\/span><i><span style=\"font-weight: 400;\">reason<\/span><\/i><span style=\"font-weight: 400;\"> about what they see. Robots with strong perception can detect obstacles, recognize objects, or interpret human gestures. Decision layers use that information to choose the best action, not just replay a fixed instruction.\u00a0<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>Traditional Control Systems:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Most industrial robots use simple control software. They follow precise, predefined motions without interpretation or context. These systems are reliable but lack adaptive autonomy. Contrast that with an AI stack where the robot continually updates its plan based on new data. Perception and reasoning are what separate reactive machines from genuinely intelligent robots.\u00a0<\/span><\/p>\n<h2><span data-text-color=\"secondary\"><b>Current State of Humanoid Robotics:<\/b><\/span><\/h2>\n<p><span style=\"font-weight: 400;\">Humanoid robots are gaining attention in research and demos around the world. You can see them at global competitions like the World Humanoid Robot Games, where more than 500 robots competed from teams in 16 countries. These robots can run, climb, and even box in front of crowds. Yet most of them still need human help for repairs or resets when something goes wrong.\u00a0<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>Progress and Limits:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Researchers have made real progress in basic movement, balance, and perception. Modern robots can handle simple grasps and obstacle navigation that would have been impossible a decade ago. However, most prototypes still struggle with real tasks outside controlled environments. <\/span><span style=\"font-weight: 400;\">They typically operate for only a few hours on a battery charge, far less than an 8 to 12 hour production shift you expect on a factory floor.\u00a0<sup>[2]<\/sup><\/span><\/p>\n<h3><span data-text-color=\"success\"><b>AI in Humanoids Today:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Technologies like computer vision and natural language processing do exist in humanoids, and some use vision models or simple language systems. But these capabilities are still limited in scope and often cannot handle complex interactions or unpredictable real\u2011world settings without human supervision or precise conditions.\u00a0<\/span><\/p>\n<h2><span data-text-color=\"secondary\"><b>Robots With Limited or Partial AI Stacks:<\/b><\/span><\/h2>\n<p><span style=\"font-weight: 400;\">Most humanoid robots that you see today use some level of artificial intelligence, but it is often focused on interaction rather than full autonomy. These robots illustrate what AI can do in limited scopes and where current technology still falls short for real industrial or general\u2011purpose use.<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>Sophia:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\"><strong><a href=\"https:\/\/qviro.com\/product\/hanson-robotics\/sophia\">Sophia<\/a><\/strong> is one of the most famous humanoid robots. She can mimic facial expressions and hold simple conversations using preprogrammed and machine\u2011assisted speech patterns. However, her responses are largely scripted and tied to predefined topics or interaction patterns rather than deep autonomous reasoning in real environments. She is built more for social engagement and research into human\u2011robot interaction than autonomous task execution on a work floor.\u00a0<\/span><\/p>\n<p><iframe loading=\"lazy\" title=\"Sophia the Robot by Hanson Robotics\" width=\"1020\" height=\"574\" src=\"https:\/\/www.youtube.com\/embed\/BhU9hOo5Cuc?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/p>\n<p style=\"text-align: center;\"><em>This video shows Sophia, one of the most recognised social humanoid robots exhibiting facial expressions and conversation in public settings. It helps illustrate interaction-focused AI rather than full autonomy.<\/em><\/p>\n<h3><span data-text-color=\"success\"><b>Nadine and Similar Social Robots:<\/b><\/span><\/h3>\n<p>Social humanoids like Nadine focus on human interaction rather than industrial performance. The robot can recognise people it has met before and recall stored information about them. It also uses speech recognition and gestures to maintain natural conversation flow. These features make it effective for reception areas, education, or public engagement. However, this type of AI is built for controlled social scenarios and demonstrations. It is not designed to handle complex tasks such as assembly, inspection, or material handling on a production line.<\/p>\n<h3><span data-text-color=\"success\"><b>Research Platforms:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Research platforms such as iCub are widely used for experimentation in perception and cognition. These robots help engineers and scientists test new algorithms in controlled settings, but they are not ready for robust deployment in everyday industrial operations. Studies show that many such robots only achieve partial autonomy and still require significant human supervision.\u00a0<\/span><\/p>\n<h2><span data-text-color=\"secondary\"><b>Humanoids With AI Stacks:<\/b><\/span><\/h2>\n<p><span style=\"font-weight: 400;\">In the broader humanoid landscape a few robots go beyond traditional control and embed stronger AI capabilities. These are not perfect autonomous machines yet but they illustrate where the technology is heading. This section shows real examples that use AI for perception, reasoning, and adaptive behaviour in ways that most robots do not.<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>Tesla Optimus:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\"><strong><a href=\"https:\/\/qviro.com\/product\/tesla\/optimus\">Tesla\u2019s Optimus<\/a><\/strong> is designed with an integrated AI stack for real\u2011world tasks. According to the official Tesla AI page, the robot\u2019s software aims to combine perception, motion planning, and interaction in unstructured environments. The company is hiring engineers in deep learning, planning and control to build these layers into Optimus. This approach pushes the robot past scripted motions toward adaptive autonomy.\u00a0<\/span><\/p>\n<p><iframe loading=\"lazy\" title=\"Optimus - Gen 2 | Tesla\" width=\"1020\" height=\"574\" src=\"https:\/\/www.youtube.com\/embed\/cpraXaw7dyc?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/p>\n<p style=\"text-align: center;\"><em>This official Tesla video shows the Optimus humanoid walking and performing tasks. It helps illustrate where AI integration is advancing beyond scripted interaction. <\/em><\/p>\n<h3><span data-text-color=\"success\"><b>Figure 03:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Figure AI\u2019s Figure 03 uses a proprietary Helix AI system designed to interpret sensor data and support reasoning for tasks at home or work. The robot features an advanced sensory suite engineered for vision\u2011language\u2011action integration, which helps it perceive its surroundings and make task decisions.\u00a0<\/span><\/p>\n<p><iframe loading=\"lazy\" title=\"Introducing Figure 03\" width=\"1020\" height=\"574\" src=\"https:\/\/www.youtube.com\/embed\/Eu5mYMavctM?feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe><\/p>\n<p style=\"text-align: center;\"><em>This video showcases Figure 03, an example of a humanoid with a stronger AI stack for perception and reasoning, useful as a contrast to limited autonomy robots.<\/em><\/p>\n<h3><span data-text-color=\"success\"><b>1X NEO:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">The <strong><a href=\"https:\/\/qviro.com\/product\/1x\/neo-gamma\">1X NEO<\/a><\/strong> humanoid is built with AI layers that enhance perception and navigation. It uses visual input and built-in connectivity to update behaviour based on new information, and the company plans to extend autonomous learning with world models that let the robot learn directly from its own footage.\u00a0<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>Social Interaction Robots:<\/b><\/span><\/h3>\n<p>Robots like <strong><a href=\"https:\/\/qviro.com\/product\/engineered-arts\/ameca\">Ameca<\/a><\/strong> demonstrate AI in human interaction through voice recognition and computer vision. While not general\u2011purpose task robots, they show how AI can be layered for perception and dialogue before being expanded into broader autonomy.<\/p>\n<p>These examples are steps toward full AI stacks but still require more development before achieving the kind of autonomy you might expect in complex industrial environments.<\/p>\n<h2><span data-text-color=\"secondary\"><b>Engineering and AI Challenges That Limit Full AI Stacks:<\/b><\/span><\/h2>\n<p><span style=\"font-weight: 400;\">Building full AI stacks for humanoid robots is not just a software problem. You need solutions for hard engineering issues that affect reliability, safety, and performance. These challenges explain why most humanoids use traditional control systems rather than end\u2011to\u2011end AI stacks.<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>1. Physical Complexity and Control:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Humanoid robots must balance, move, and coordinate many joints to perform tasks. This is difficult even with fixed routines. Most production robots rely on classical motion control systems that you can predict and tune precisely. According to research, real\u2011time control and stability remain top challenges for dynamic locomotion and manipulation.\u00a0<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>2. Data Scarcity:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">AI models need large datasets to learn. Language models train on trillions of words. In contrast there is far less motion and interaction data for humanoid robots. Simulations help but they cannot fully mimic real physics and contact forces, which remain a major gap in training autonomous behaviours.<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>3. Compute and Power Constraints:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">AI stacks require heavy processing and high energy use. Humanoid robots run on batteries with limited power. This makes it hard to run advanced perception and planning in real time without overheating or draining power quickly.<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>4. Safety and Reliability:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Industrial automation demands predictable performance. Autonomous AI that changes behaviour based on context can produce unexpected actions. Engineers prefer systems with clear limits and fail\u2011safe modes. This prioritizes reliability over adaptive autonomy in most current humanoids.<\/span><\/p>\n<h2><span data-text-color=\"secondary\"><b>Emerging Trends: Towards Full AI Stacks<\/b><\/span><\/h2>\n<p>The robotics industry is beginning to build systems that edge closer to full AI stacks. One key trend is the use of world models and foundation models that help robots understand and predict physical environments. These models aim to integrate perception, reasoning, and planning into a unified framework, which is a step toward autonomy rather than fixed routines. Researchers have developed open world models that forecast future observations and help robots anticipate outcomes based on actions. This research suggests a path for embodied AI that goes beyond traditional control.<\/p>\n<h3><span data-text-color=\"success\"><b>Foundation Models in Robotics:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Foundation models are powerful AI systems trained on large datasets and used across multiple tasks, including vision and language. In robotics, they show potential to improve perception, decision\u2011making, and control by transferring learned knowledge into autonomous behaviour. However, the field still faces challenges with data scarcity, safety guarantees, and real\u2011time execution.<sup>[3]<\/sup><\/span><span style=\"font-weight: 400;\">\u00a0<\/span><\/p>\n<h3><span data-text-color=\"success\"><b>Progress But Limited Adoption:<\/b><\/span><\/h3>\n<p><span style=\"font-weight: 400;\">Industry momentum is visible, and national initiatives aim to support full\u2011stack humanoid ecosystems. Yet production\u2011ready AI stacks remain rare outside of research prototypes and experimental platforms. Progressing from laboratory models to reliable, industrial\u2011grade AI stacks continues to require breakthroughs in hardware, data, and safety. <sup>[4] <\/sup><\/span><\/p>\n<h2><span data-text-color=\"secondary\"><b>Conclusion:<\/b><\/span><\/h2>\n<p><span style=\"font-weight: 400;\">Most humanoid robots today do not have full AI stacks. The reasons are practical and engineering\u2011based. Balancing and coordinating multiple joints is complex. Data for training autonomous behaviours is limited compared to what large AI models use. Processing power and battery life constrain what can run on a mobile robot. Safety and reliability requirements make unpredictable AI behaviour risky in real environments. For now, engineers rely on a mix of traditional control systems and selective AI modules for perception or simple decision\u2011making. This approach is the practical reality. Over time, as hardware, data, and algorithms improve, humanoid robots will become smarter and more capable.<\/span><\/p>\n<h2><span data-text-color=\"secondary\"><strong>References:<\/strong><\/span><\/h2>\n<ol>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">IBM. <\/span><i><span style=\"font-weight: 400;\">AI stack in robotics<\/span><\/i><span style=\"font-weight: 400;\">. Retrieved on 6 January 2026, from <\/span><a href=\"https:\/\/www.ibm.com\/think\/topics\/ai-stack?utm_source=chatgpt.com\"><span style=\"font-weight: 400;\">https:\/\/www.ibm.com\/think\/topics\/ai-stack<\/span><\/a><span style=\"font-weight: 400;\">?<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">McKinsey &amp; Company. <\/span><i><span style=\"font-weight: 400;\">Humanoid robots crossing the chasm from concept to commercial reality<\/span><\/i><span style=\"font-weight: 400;\">. Retrieved on 6 January 2026, from <\/span><a href=\"https:\/\/www.mckinsey.com\/industries\/industrials\/our-insights\/humanoid-robots-crossing-the-chasm-from-concept-to-commercial-reality?utm_source=chatgpt.com\"><span style=\"font-weight: 400;\">https:\/\/www.mckinsey.com\/industries\/industrials\/our-insights\/humanoid-robots-crossing-the-chasm-from-concept-to-commercial-reality<\/span><\/a><span style=\"font-weight: 400;\">?<\/span><\/li>\n<li style=\"font-weight: 400;\" aria-level=\"1\"><span style=\"font-weight: 400;\">IBM. <\/span><i><span style=\"font-weight: 400;\">Foundation models<\/span><\/i><span style=\"font-weight: 400;\">. Retrieved on 6 January 2026, from <\/span><a href=\"https:\/\/www.ibm.com\/think\/topics\/foundation-models?utm_source=chatgpt.com\"><span style=\"font-weight: 400;\">https:\/\/www.ibm.com\/think\/topics\/foundation-models<\/span><\/a><span style=\"font-weight: 400;\">?<\/span><\/li>\n<li aria-level=\"1\"><span style=\"font-weight: 400;\">McKinsey &amp; Company (Germany). <\/span><i><span style=\"font-weight: 400;\">Humanoid robots crossing the chasm from concept to commercial reality<\/span><\/i><span style=\"font-weight: 400;\">. Retrieved on 6 January 2026, from <\/span><a href=\"https:\/\/www.mckinsey.de\/industries\/industrials\/our-insights\/humanoid-robots-crossing-the-chasm-from-concept-to-commercial-reality?utm_source=chatgpt.com\"><span style=\"font-weight: 400;\">https:\/\/www.mckinsey.de\/industries\/industrials\/our-insights\/humanoid-robots-crossing-the-chasm-from-concept-to-commercial-reality<\/span><\/a><span style=\"font-weight: 400;\">?<\/span><\/li>\n<\/ol>\n\n  \n    <div class=\"row large-columns-3 medium-columns- small-columns-1 slider row-slider slider-nav-reveal slider-nav-push\"  data-flickity-options='{\"imagesLoaded\": true, \"groupCells\": \"100%\", \"dragThreshold\" : 5, \"cellAlign\": \"left\",\"wrapAround\": true,\"prevNextButtons\": true,\"percentPosition\": true,\"pageDots\": false, \"rightToLeft\": false, \"autoPlay\" : false}' >\n\n  <div class=\"col post-item\" >\n\t\t\t<div class=\"col-inner\">\n\t\t\t\t<div class=\"box box-text-bottom box-blog-post has-hover\">\n            \t\t\t\t\t<div class=\"box-image\" >\n  \t\t\t\t\t\t<div class=\"image-cover\" style=\"padding-top:200px;\">\n\t\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/mobile-robot-battery\/\" class=\"plain\" aria-label=\"Mobile Robot Battery: Swappable vs. Fast-Charge\">\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"600\" height=\"400\" src=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/04\/Mobile-Robot-Batteries-Swappable-vs.-Fast-Charge-600x400.webp\" class=\"attachment-medium size-medium wp-post-image\" alt=\"\" srcset=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/04\/Mobile-Robot-Batteries-Swappable-vs.-Fast-Charge-600x400.webp 600w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/04\/Mobile-Robot-Batteries-Swappable-vs.-Fast-Charge-768x512.webp 768w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/04\/Mobile-Robot-Batteries-Swappable-vs.-Fast-Charge.webp 1000w\" sizes=\"auto, (max-width: 600px) 100vw, 600px\" \/>\t\t\t\t\t\t\t<\/a>\n  \t\t\t\t\t\t\t  \t\t\t\t\t\t\t  \t\t\t\t\t\t<\/div>\n  \t\t\t\t\t\t  \t\t\t\t\t<\/div>\n          \t\t\t\t\t<div class=\"box-text text-center\" >\n\t\t\t\t\t<div class=\"box-text-inner blog-post-inner\">\n\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t<h5 class=\"post-title is-large \">\n\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/mobile-robot-battery\/\" class=\"plain\">Mobile Robot Battery: Swappable vs. Fast-Charge<\/a>\n\t\t\t\t\t<\/h5>\n\t\t\t\t\t\t\t\t\t\t<div class=\"is-divider\"><\/div>\n\t\t\t\t\t\t\t\t\t\t<p class=\"from_the_blog_excerpt \">Every minute a mobile robot spends off the floor is a minute of lost productivity. [...]\t\t\t\t\t<\/p>\n\t\t\t\t\t                    \n\t\t\t\t\t\n\t\t\t\t\t\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<div class=\"badge absolute top post-date badge-square\">\n\t\t\t\t\t\t\t<div class=\"badge-inner\">\n\t\t\t\t\t\t\t\t<span class=\"post-date-day\">13<\/span><br>\n\t\t\t\t\t\t\t\t<span class=\"post-date-month is-xsmall\">Apr<\/span>\n\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t<\/div>\n\t\t<\/div><div class=\"col post-item\" >\n\t\t\t<div class=\"col-inner\">\n\t\t\t\t<div class=\"box box-text-bottom box-blog-post has-hover\">\n            \t\t\t\t\t<div class=\"box-image\" >\n  \t\t\t\t\t\t<div class=\"image-cover\" style=\"padding-top:200px;\">\n\t\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/force-limiting-cobots\/\" class=\"plain\" aria-label=\"What is Force Limiting in Cobots?\">\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"400\" height=\"400\" src=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/What-is-Force-Limiting-in-Cobots-400x400.webp\" class=\"attachment-medium size-medium wp-post-image\" alt=\"\" srcset=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/What-is-Force-Limiting-in-Cobots-400x400.webp 400w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/What-is-Force-Limiting-in-Cobots-800x800.webp 800w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/What-is-Force-Limiting-in-Cobots-280x280.webp 280w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/What-is-Force-Limiting-in-Cobots-768x768.webp 768w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/What-is-Force-Limiting-in-Cobots.webp 1475w\" sizes=\"auto, (max-width: 400px) 100vw, 400px\" \/>\t\t\t\t\t\t\t<\/a>\n  \t\t\t\t\t\t\t  \t\t\t\t\t\t\t  \t\t\t\t\t\t<\/div>\n  \t\t\t\t\t\t  \t\t\t\t\t<\/div>\n          \t\t\t\t\t<div class=\"box-text text-center\" >\n\t\t\t\t\t<div class=\"box-text-inner blog-post-inner\">\n\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t<h5 class=\"post-title is-large \">\n\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/force-limiting-cobots\/\" class=\"plain\">What is Force Limiting in Cobots?<\/a>\n\t\t\t\t\t<\/h5>\n\t\t\t\t\t\t\t\t\t\t<div class=\"is-divider\"><\/div>\n\t\t\t\t\t\t\t\t\t\t<p class=\"from_the_blog_excerpt \">Collaborative robots, or cobots, operate in shared human workspaces. This capability requires robust safety mechanisms. [...]\t\t\t\t\t<\/p>\n\t\t\t\t\t                    \n\t\t\t\t\t\n\t\t\t\t\t\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<div class=\"badge absolute top post-date badge-square\">\n\t\t\t\t\t\t\t<div class=\"badge-inner\">\n\t\t\t\t\t\t\t\t<span class=\"post-date-day\">26<\/span><br>\n\t\t\t\t\t\t\t\t<span class=\"post-date-month is-xsmall\">Mar<\/span>\n\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t<\/div>\n\t\t<\/div><div class=\"col post-item\" >\n\t\t\t<div class=\"col-inner\">\n\t\t\t\t<div class=\"box box-text-bottom box-blog-post has-hover\">\n            \t\t\t\t\t<div class=\"box-image\" >\n  \t\t\t\t\t\t<div class=\"image-cover\" style=\"padding-top:200px;\">\n\t\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/structured-light-time-of-flight-cameras\/\" class=\"plain\" aria-label=\"Structured Light vs. Time-of-Flight Cameras\">\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"599\" height=\"400\" src=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/Structured-Light-vs-Time-of-Flight-Cameras-599x400.webp\" class=\"attachment-medium size-medium wp-post-image\" alt=\"\" srcset=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/Structured-Light-vs-Time-of-Flight-Cameras-599x400.webp 599w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/Structured-Light-vs-Time-of-Flight-Cameras-1199x800.webp 1199w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/Structured-Light-vs-Time-of-Flight-Cameras-768x513.webp 768w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/Structured-Light-vs-Time-of-Flight-Cameras-1536x1025.webp 1536w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/03\/Structured-Light-vs-Time-of-Flight-Cameras.webp 1681w\" sizes=\"auto, (max-width: 599px) 100vw, 599px\" \/>\t\t\t\t\t\t\t<\/a>\n  \t\t\t\t\t\t\t  \t\t\t\t\t\t\t  \t\t\t\t\t\t<\/div>\n  \t\t\t\t\t\t  \t\t\t\t\t<\/div>\n          \t\t\t\t\t<div class=\"box-text text-center\" >\n\t\t\t\t\t<div class=\"box-text-inner blog-post-inner\">\n\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t<h5 class=\"post-title is-large \">\n\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/structured-light-time-of-flight-cameras\/\" class=\"plain\">Structured Light vs. Time-of-Flight Cameras<\/a>\n\t\t\t\t\t<\/h5>\n\t\t\t\t\t\t\t\t\t\t<div class=\"is-divider\"><\/div>\n\t\t\t\t\t\t\t\t\t\t<p class=\"from_the_blog_excerpt \">Depth sensing enables robots to perceive the three-dimensional structure of their surroundings. It supports tasks [...]\t\t\t\t\t<\/p>\n\t\t\t\t\t                    \n\t\t\t\t\t\n\t\t\t\t\t<div class=\"col qviro-ad-inject\" style=\"text-align: center;\"><div class='code-block code-block-3' style='margin: 8px 0; clear: both;'>\n<div id='div-gpt-ad-listing-300x250' style=\"min-width: 300px; min-height: 250px; width: 300px; margin: 0 auto;\">\n  <script>\n    googletag.cmd.push(function() {\n      googletag.defineSlot('\/4250\/qviro\/blog\/listing', [300, 250], 'div-gpt-ad-listing-300x250').addService(googletag.pubads());\n      googletag.display('div-gpt-ad-listing-300x250');\n    });\n  <\/script>\n<\/div><\/div>\n<\/div><script>\n\t\t(function(){\n\t\t\tvar ad = document.currentScript.previousElementSibling;\n\t\t\tvar postItem = ad.closest(\".post-item\");\n\t\t\tif(postItem && ad) {\n\t\t\t\tpostItem.parentNode.insertBefore(ad, postItem.nextSibling);\n\t\t\t}\n\n\t\t\t\/\/ Listen for GPT render events to hide container if no ad serves\n\t\t\t\/\/ Skip collapse if Google Publisher Console is active (for debugging)\n\t\t\tvar isDebugging = window.location.search.includes(\"googfc\");\n\n\t\t\tif (!isDebugging && typeof googletag !== \"undefined\") {\n\t\t\t\tgoogletag.cmd.push(function() {\n\t\t\t\t\tgoogletag.pubads().addEventListener(\"slotRenderEnded\", function(event) {\n\t\t\t\t\t\tvar slotId = event.slot.getSlotElementId();\n\t\t\t\t\t\t\/\/ Check if this is one of our listing ads (block 3 or 4)\n\t\t\t\t\t\tif (slotId === \"div-gpt-ad-listing-300x250\" || slotId === \"div-gpt-ad-listing-320x50\") {\n\t\t\t\t\t\t\tif (event.isEmpty) {\n\t\t\t\t\t\t\t\tad.style.display = \"none\";\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t}\n\t\t\t\t\t});\n\t\t\t\t});\n\t\t\t}\n\t\t})();\n\t\t<\/script>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<div class=\"badge absolute top post-date badge-square\">\n\t\t\t\t\t\t\t<div class=\"badge-inner\">\n\t\t\t\t\t\t\t\t<span class=\"post-date-day\">15<\/span><br>\n\t\t\t\t\t\t\t\t<span class=\"post-date-month is-xsmall\">Mar<\/span>\n\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t<\/div>\n\t\t<\/div><div class=\"col post-item\" >\n\t\t\t<div class=\"col-inner\">\n\t\t\t\t<div class=\"box box-text-bottom box-blog-post has-hover\">\n            \t\t\t\t\t<div class=\"box-image\" >\n  \t\t\t\t\t\t<div class=\"image-cover\" style=\"padding-top:200px;\">\n\t\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/human-machine-interface\/\" class=\"plain\" aria-label=\"Human Machine Interface Explained\">\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"400\" height=\"400\" src=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Human-Machine-Interface-Explained-400x400.webp\" class=\"attachment-medium size-medium wp-post-image\" alt=\"\" srcset=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Human-Machine-Interface-Explained-400x400.webp 400w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Human-Machine-Interface-Explained-800x800.webp 800w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Human-Machine-Interface-Explained-280x280.webp 280w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Human-Machine-Interface-Explained-768x768.webp 768w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Human-Machine-Interface-Explained.webp 1047w\" sizes=\"auto, (max-width: 400px) 100vw, 400px\" \/>\t\t\t\t\t\t\t<\/a>\n  \t\t\t\t\t\t\t  \t\t\t\t\t\t\t  \t\t\t\t\t\t<\/div>\n  \t\t\t\t\t\t  \t\t\t\t\t<\/div>\n          \t\t\t\t\t<div class=\"box-text text-center\" >\n\t\t\t\t\t<div class=\"box-text-inner blog-post-inner\">\n\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t<h5 class=\"post-title is-large \">\n\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/human-machine-interface\/\" class=\"plain\">Human Machine Interface Explained<\/a>\n\t\t\t\t\t<\/h5>\n\t\t\t\t\t\t\t\t\t\t<div class=\"is-divider\"><\/div>\n\t\t\t\t\t\t\t\t\t\t<p class=\"from_the_blog_excerpt \">If you have automated equipment, you interact with a human machine interface on a daily [...]\t\t\t\t\t<\/p>\n\t\t\t\t\t                    \n\t\t\t\t\t\n\t\t\t\t\t\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<div class=\"badge absolute top post-date badge-square\">\n\t\t\t\t\t\t\t<div class=\"badge-inner\">\n\t\t\t\t\t\t\t\t<span class=\"post-date-day\">26<\/span><br>\n\t\t\t\t\t\t\t\t<span class=\"post-date-month is-xsmall\">Feb<\/span>\n\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t<\/div>\n\t\t<\/div><div class=\"col post-item\" >\n\t\t\t<div class=\"col-inner\">\n\t\t\t\t<div class=\"box box-text-bottom box-blog-post has-hover\">\n            \t\t\t\t\t<div class=\"box-image\" >\n  \t\t\t\t\t\t<div class=\"image-cover\" style=\"padding-top:200px;\">\n\t\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/why-most-humanoid-robots-lack-ai-stacks\/\" class=\"plain\" aria-label=\"Why Most Humanoid Robots Lack AI Stacks\">\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"749\" height=\"400\" src=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Why-Most-Humanoid-Robots-Lack-AI-Stacks-749x400.webp\" class=\"attachment-medium size-medium wp-post-image\" alt=\"\" srcset=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Why-Most-Humanoid-Robots-Lack-AI-Stacks-749x400.webp 749w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Why-Most-Humanoid-Robots-Lack-AI-Stacks-1400x748.webp 1400w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Why-Most-Humanoid-Robots-Lack-AI-Stacks-768x410.webp 768w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/02\/Why-Most-Humanoid-Robots-Lack-AI-Stacks.webp 1483w\" sizes=\"auto, (max-width: 749px) 100vw, 749px\" \/>\t\t\t\t\t\t\t<\/a>\n  \t\t\t\t\t\t\t  \t\t\t\t\t\t\t  \t\t\t\t\t\t<\/div>\n  \t\t\t\t\t\t  \t\t\t\t\t<\/div>\n          \t\t\t\t\t<div class=\"box-text text-center\" >\n\t\t\t\t\t<div class=\"box-text-inner blog-post-inner\">\n\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t<h5 class=\"post-title is-large \">\n\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/why-most-humanoid-robots-lack-ai-stacks\/\" class=\"plain\">Why Most Humanoid Robots Lack AI Stacks<\/a>\n\t\t\t\t\t<\/h5>\n\t\t\t\t\t\t\t\t\t\t<div class=\"is-divider\"><\/div>\n\t\t\t\t\t\t\t\t\t\t<p class=\"from_the_blog_excerpt \">Humanoid robots are robots built to look and move like humans. They usually have two [...]\t\t\t\t\t<\/p>\n\t\t\t\t\t                    \n\t\t\t\t\t\n\t\t\t\t\t\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<div class=\"badge absolute top post-date badge-square\">\n\t\t\t\t\t\t\t<div class=\"badge-inner\">\n\t\t\t\t\t\t\t\t<span class=\"post-date-day\">16<\/span><br>\n\t\t\t\t\t\t\t\t<span class=\"post-date-month is-xsmall\">Feb<\/span>\n\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t<\/div>\n\t\t<\/div><div class=\"col post-item\" >\n\t\t\t<div class=\"col-inner\">\n\t\t\t\t<div class=\"box box-text-bottom box-blog-post has-hover\">\n            \t\t\t\t\t<div class=\"box-image\" >\n  \t\t\t\t\t\t<div class=\"image-cover\" style=\"padding-top:200px;\">\n\t\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/january-2026-humanoid-robot-launches\/\" class=\"plain\" aria-label=\"January 2026 Humanoid Robot Launches\">\n\t\t\t\t\t\t\t\t<img loading=\"lazy\" decoding=\"async\" width=\"600\" height=\"400\" src=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/01\/January-2026-Humanoid-Robot-Launches-600x400.webp\" class=\"attachment-medium size-medium wp-post-image\" alt=\"\" srcset=\"https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/01\/January-2026-Humanoid-Robot-Launches-600x400.webp 600w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/01\/January-2026-Humanoid-Robot-Launches-1200x800.webp 1200w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/01\/January-2026-Humanoid-Robot-Launches-768x512.webp 768w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/01\/January-2026-Humanoid-Robot-Launches-1536x1024.webp 1536w, https:\/\/qviro.com\/blog\/wp-content\/uploads\/2026\/01\/January-2026-Humanoid-Robot-Launches-2048x1366.webp 2048w\" sizes=\"auto, (max-width: 600px) 100vw, 600px\" \/>\t\t\t\t\t\t\t<\/a>\n  \t\t\t\t\t\t\t  \t\t\t\t\t\t\t  \t\t\t\t\t\t<\/div>\n  \t\t\t\t\t\t  \t\t\t\t\t<\/div>\n          \t\t\t\t\t<div class=\"box-text text-center\" >\n\t\t\t\t\t<div class=\"box-text-inner blog-post-inner\">\n\n\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t<h5 class=\"post-title is-large \">\n\t\t\t\t\t\t<a href=\"https:\/\/qviro.com\/blog\/january-2026-humanoid-robot-launches\/\" class=\"plain\">January 2026 Humanoid Robot Launches<\/a>\n\t\t\t\t\t<\/h5>\n\t\t\t\t\t\t\t\t\t\t<div class=\"is-divider\"><\/div>\n\t\t\t\t\t\t\t\t\t\t<p class=\"from_the_blog_excerpt \">January marked a major milestone for the humanoid robotics sector, as several leading manufacturers unveiled [...]\t\t\t\t\t<\/p>\n\t\t\t\t\t                    \n\t\t\t\t\t\n\t\t\t\t\t\n\t\t\t\t\t<\/div>\n\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t<div class=\"badge absolute top post-date badge-square\">\n\t\t\t\t\t\t\t<div class=\"badge-inner\">\n\t\t\t\t\t\t\t\t<span class=\"post-date-day\">23<\/span><br>\n\t\t\t\t\t\t\t\t<span class=\"post-date-month is-xsmall\">Jan<\/span>\n\t\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t<\/div>\n\t\t\t\t\t\t\t\t\t<\/div>\n\t\t\t<\/div>\n\t\t<\/div><\/div>\n","protected":false},"excerpt":{"rendered":"<p>Humanoid robots are robots built to look and move like humans. They usually have two arms, two legs, and a head. The goal is simple: let the robot work in spaces designed for people, using the same tools and workflows you already have. Today, many of these robots are marketed as AI-powered. But when you [&#8230;]\n","protected":false},"author":7,"featured_media":33521,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"om_disable_all_campaigns":false,"footnotes":""},"categories":[151],"tags":[385,377,182,155,143],"class_list":["post-33512","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-humanoid-robot","tag-1x","tag-engineered-arts","tag-figure","tag-hanson-robotics","tag-tesla"],"aioseo_notices":[],"_links":{"self":[{"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/posts\/33512","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/users\/7"}],"replies":[{"embeddable":true,"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/comments?post=33512"}],"version-history":[{"count":9,"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/posts\/33512\/revisions"}],"predecessor-version":[{"id":33523,"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/posts\/33512\/revisions\/33523"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/media\/33521"}],"wp:attachment":[{"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/media?parent=33512"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/categories?post=33512"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/qviro.com\/blog\/wp-json\/wp\/v2\/tags?post=33512"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}