{"id":589799,"date":"2026-04-17T13:20:08","date_gmt":"2026-04-17T13:20:08","guid":{"rendered":"https:\/\/www.newsbeep.com\/us\/589799\/"},"modified":"2026-04-17T13:20:08","modified_gmt":"2026-04-17T13:20:08","slug":"robots-are-learning-to-do-housework-from-videos-of-humans-doing-chores","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/us\/589799\/","title":{"rendered":"Robots are learning to do housework from videos of humans doing chores"},"content":{"rendered":"<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"NGCAUWLMLBCSJJFNQ2U544TCAM\" data-contentid=\"NGCAUWLMLBCSJJFNQ2U544TCAM\">Silicon Valley\u2019s next great leap may be built on videos of people folding laundry.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"VW9K4JCMN54MN34XBUFWY0YMAR\" data-contentid=\"VW9K4JCMN54MN34XBUFWY0YMAR\">Start-ups and entrepreneurs including <a href=\"https:\/\/www.washingtonpost.com\/technology\/2026\/03\/27\/musk-optimus-robot-physical-ai\/?itid=lk_inline_enhanced-template\" rel=\"nofollow noopener\" target=\"_blank\">Tesla CEO Elon Musk<\/a> are trying to make <a href=\"https:\/\/www.washingtonpost.com\/technology\/2025\/09\/05\/humanoid-robots-ai-agility-chatgpt\/?itid=lk_inline_enhanced-template\" rel=\"nofollow noopener\" target=\"_blank\">robots smart enough<\/a> to help with chores around the home. But adapting artificial intelligence to new tasks requires example data, like the <a href=\"https:\/\/www.washingtonpost.com\/technology\/interactive\/2023\/ai-chatbot-learning\/?itid=lk_inline_enhanced-template\" rel=\"nofollow noopener\" target=\"_blank\">online text<\/a> and <a href=\"https:\/\/www.washingtonpost.com\/technology\/interactive\/2022\/ai-image-generator\/?itid=lk_inline_enhanced-template\" rel=\"nofollow noopener\" target=\"_blank\">photos<\/a> that enabled chatbots to start generating high-quality documents and images.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"E7DDEDJDK920D6VNHX46XJ6UT8\" data-contentid=\"E7DDEDJDK920D6VNHX46XJ6UT8\">DoorDash, a food-delivery service, has joined a cottage industry of companies and researchers gathering data for the robot revolution in the form of videos of people doing tasks like folding clothes or washing dishes. Gig workers can now earn as much as $25 an hour by recording themselves doing chores for DoorDash.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"506TK4349521152C8FZ900GP4C\" data-contentid=\"506TK4349521152C8FZ900GP4C\">Here\u2019s how that video is used and why it\u2019s so valuable.<\/p>\n<p>How to train robots to fold clothes<\/p>\n<p>\t\t<img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/us\/wp-content\/uploads\/2026\/04\/2300-ROBOT-DATA-training.png\" id=\"fallback-image\" style=\"position: relative;\" class=\"g-aiImg\" alt=\"\"\/><\/p>\n<p class=\"g-pstyle1\">People fold laundry while wearing a head-mounted smartphone. The phone records video, which is processed to track movement of the head, hands and fingers.<\/p>\n<p class=\"g-pstyle1\">To help transfer the skill from human fingers to robot movement commands, AI researchers record themselves folding clothes using robotic limbs. <\/p>\n<p class=\"g-pstyle1\">To help AI algorithms translate the skill from human fingers to robot movement commands, data is also collected from expert robot operators folding clothes using remote-controlled robotic limbs.<\/p>\n<p class=\"g-pstyle1\">Researchers feed all the data into machine learning algorithms, which learn to predict from what a robot sees through its cameras and sensors what movements it should make to fold an item of clothing.<\/p>\n<p>\u00a0<\/p>\n<p class=\"g-pstyle1\">Finally, the AI model is loaded into robots that attempt to fold clothes autonomously.<\/p>\n<p class=\"g-pstyle2\">Videos from EgoMimic, EgoVerse and Nvidia<\/p>\n<p class=\"g-pstyle1\">People fold laundry while wearing a head-mounted smartphone. The phone records video, which is processed to track movement of the head, hands and fingers.<\/p>\n<p class=\"g-pstyle1\">To capture a variety of situations a robot may need to understand, people are recruited around the world to record themselves as they fold different types of clothes on different surfaces.<\/p>\n<p class=\"g-pstyle1\">To help transfer the skill from human fingers to robot movement commands, AI researchers record themselves folding clothes using robotic limbs.<\/p>\n<p class=\"g-pstyle1\">Researchers feed all the data into machine learning algorithms, which learn to predict from what a robot sees through its cameras and sensors what movements it should make to fold an item of clothing.<\/p>\n<p>\u00a0<\/p>\n<p class=\"g-pstyle1\">Finally, the AI model is loaded into robots that attempt to fold clothes autonomously.<\/p>\n<p class=\"g-pstyle2\">Videos from EgoMimic, EgoVerse and Nvidia<\/p>\n<p class=\"g-pstyle1\">People fold laundry while wearing a head-mounted smartphone. The phone records video, which is processed to track movement of the head, hands and fingers.<\/p>\n<p class=\"g-pstyle1\">To capture a variety of situations a robot may need to understand, people are recruited around the world to record themselves as they fold different types of clothes on different surfaces.<\/p>\n<p class=\"g-pstyle1\">To help transfer the skill from human fingers to robot movement commands, AI researchers record themselves folding clothes using robotic limbs.<\/p>\n<p class=\"g-pstyle1\">Researchers feed the data into a machine learning process, where an AI model learns to predict what movements to take next to fold a shirt.<\/p>\n<p>\u00a0<\/p>\n<p class=\"g-pstyle1\">Finally, the AI model is loaded into robots that attempt to fold clothes autonomously.<\/p>\n<p class=\"g-pstyle3\">Videos from EgoMimic, EgoVerse and Nvidia<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"Z2BTU9KGD14KZ7Z6Q944HVAQ6C\" data-contentid=\"Z2BTU9KGD14KZ7Z6Q944HVAQ6C\">The household chores data grab is a bet on what AI insiders call scaling laws. Researchers have found that AI models for working with text or images get progressively better the more data they were trained on, and researchers hope the same is true for robotics.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"NCM4BKQYHN3G776NP2N3QY5NV4\" data-contentid=\"NCM4BKQYHN3G776NP2N3QY5NV4\">\u201cThere is evidence that a lot of data would help\u201d robots do more complex tasks, said Ken Goldberg, a roboticist and distinguished chair of engineering at the University of California at Berkeley.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"6UYMVHJ3QJC4XMYZCSX4VJAOFQ\" data-contentid=\"6UYMVHJ3QJC4XMYZCSX4VJAOFQ\">But unlike for chatbots, there isn\u2019t an easy place to get oceans of relevant data. \u201cThere\u2019s no internet for robot data,\u201d Goldberg said.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"GQ4OJFZCWVBSJJEMLBWXKNTVXA\" data-contentid=\"GQ4OJFZCWVBSJJEMLBWXKNTVXA\">Chatbots learn to generate coherent sentences by analyzing human-written text, raw material that is readily available from the web, books or numerous other sources.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"WK52BZ5BABBF7CARCYW62MCX5Q\" data-contentid=\"WK52BZ5BABBF7CARCYW62MCX5Q\">Training robot control software is more complicated. To take on household chores, a robot needs to decipher data from its sensors, predict which actions will achieve a goal like folding a shirt, and send commands to limbs and grippers to make the appropriate motions. There\u2019s no ready-made repository of data demonstrating how to do that. Even videos of people doing chores don\u2019t have all the elements needed.<\/p>\n<p>Size of AI training datasets<\/p>\n<p>\t\t<img decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/us\/wp-content\/uploads\/2026\/04\/2300-ROBOT-DATA-scale.png\" id=\"fallback-image\" style=\"position: relative;\" class=\"g-aiImg\" alt=\"\"\/><\/p>\n<p>\t\t<img decoding=\"async\" id=\"g-ROBOT-DATA-scale-small-img\" class=\"g-ROBOT-DATA-scale-small-img g-aiImg\" alt=\"\" src=\"https:\/\/gfx-data.news-engineering.aws.wapo.pub\/ai2html\/ROBOT-DATA\/RWZ4AG6ZIBGBTLVCWMLG4NXR6Y\/ROBOT-DATA-scale-small.png?v=15\" bad-src=\"data:image\/gif;base64,R0lGODlhCgAKAIAAAB8fHwAAACH5BAEAAAAALAAAAAAKAAoAAAIIhI+py+0PYysAOw==\"\/><\/p>\n<p class=\"g-pstyle0\">Sizes of AI training datasets<\/p>\n<p class=\"g-pstyle1\">= 10 years of human effort<\/p>\n<p class=\"g-pstyle3\">~5 years to watch<\/p>\n<p class=\"g-pstyle3\">(\u00bd square)<\/p>\n<p class=\"g-pstyle4\">Dataset for training chatbots<\/p>\n<p class=\"g-pstyle6\">The chatbot estimate assumes a 1.5 trillion token dataset, at 1.33 tokens per word and a reading pace of 238 words per minute.<\/p>\n<p class=\"g-pstyle7\">Source: Estimate by robotics researcher Kevin Black of UC-Berkeley and Physical Intelligence<\/p>\n<p>\t\t<img decoding=\"async\" id=\"g-ROBOT-DATA-scale-medium-img\" class=\"g-ROBOT-DATA-scale-medium-img g-aiImg\" alt=\"\" src=\"https:\/\/gfx-data.news-engineering.aws.wapo.pub\/ai2html\/ROBOT-DATA\/RWZ4AG6ZIBGBTLVCWMLG4NXR6Y\/ROBOT-DATA-scale-medium.png?v=15\" bad-src=\"data:image\/gif;base64,R0lGODlhCgAKAIAAAB8fHwAAACH5BAEAAAAALAAAAAAKAAoAAAIIhI+py+0PYysAOw==\"\/><\/p>\n<p class=\"g-pstyle0\">Sizes of AI training datasets<\/p>\n<p class=\"g-pstyle1\">= 10 years of human effort<\/p>\n<p class=\"g-pstyle2\">Dataset for training chatbots<\/p>\n<p class=\"g-pstyle1\">~5 years to watch (\u00bd square)<\/p>\n<p class=\"g-pstyle3\">The chatbot estimate assumes a 1.5 trillion token dataset, at 1.33 tokens per word and a reading pace of 238 words per minute.<\/p>\n<p class=\"g-pstyle4\">Source: Estimate by robotics researcher Kevin Black of UC-Berkeley and Physical Intelligence<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"7AJB3NBUARGPVIPWHOEQJSQLGU\" data-contentid=\"7AJB3NBUARGPVIPWHOEQJSQLGU\">One way to gather training material is to record data while humans manually operate robots. \u201cRobot teleoperation data is probably considered the highest quality of data,\u201d because it includes robot motion commands, said Simar Kareer, a robotics researcher at Georgia Tech who helped pioneer training robots on human videos.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"WTRHCPT6UVHV7A3WBYDECJHDAE\" data-contentid=\"WTRHCPT6UVHV7A3WBYDECJHDAE\">But \u201cit\u2019s just the most expensive to collect,\u201d Kareer said, because you have to pay people to operate an expensive robot, and \u201cthe person is completing these tasks much, much slower than they would if they were using their own hands.\u201d<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"TBQ6EVME2529B6BQ3TK5RBAWB8\" data-contentid=\"TBQ6EVME2529B6BQ3TK5RBAWB8\">Kareer is working to show that a large collection of cheaper human video data can provide AI a baseline understanding of how to do tasks that can be refined with a smaller pool of expensive teleoperation data to teach software how to make specific robotic actions.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"JU141DZ7U94TF71PUA7A3ZQ88G\" data-contentid=\"JU141DZ7U94TF71PUA7A3ZQ88G\">Other researchers and companies are trying different tactics to reduce the cost of gathering the training data needed for a robot revolution.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"2XDHMGXW3H3CBD9M1A9BMQ4B6W\" data-contentid=\"2XDHMGXW3H3CBD9M1A9BMQ4B6W\"><a href=\"https:\/\/www.sunday.ai\/technology?itid=lk_inline_enhanced-template\" rel=\"nofollow noopener\" target=\"_blank\">One<\/a> is to give humans a handheld version of a robot gripper to make it easier and quicker to demonstrate tasks in a way easily translated to robot control software. <a href=\"https:\/\/www.mimicrobotics.com\/technology?itid=lk_inline_enhanced-template\" rel=\"nofollow noopener\" target=\"_blank\">Others<\/a> build robots to be as similar to humans as possible. If a machine has the same number of fingers and joints as humans, the thinking goes, it will be easier for AI software to transfer skills from human videos to robots. <a href=\"https:\/\/www.nvidia.com\/en-us\/use-cases\/robotics-simulation\/?itid=lk_inline_enhanced-template\" rel=\"nofollow noopener\" target=\"_blank\">Another<\/a> idea is to let robots experiment and learn in a simulated environment, like a video game, before transferring the control software onto real robots.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"XECF6AETAD32NCY6402EDV41TG\" data-contentid=\"XECF6AETAD32NCY6402EDV41TG\">Ultimately, the best data for making robots better at folding your clothes will come after they begin to be deployed and doing real tasks in the world. But it\u2019s not clear how soon that will be possible.<\/p>\n<p data-component=\"Text\" class=\"wpds-c-foYyTm wpds-c-foYyTm-idAgRsb-css component-text\" id=\"3BFMJN5EVH2MBEXURKEWZ9YNNM\" data-contentid=\"3BFMJN5EVH2MBEXURKEWZ9YNNM\">How long until a robot can do your laundry? \u201cMaybe in two years, three, five, 10, 20,\u201d Goldberg said. \u201cOr longer.\u201d<\/p>\n","protected":false},"excerpt":{"rendered":"Silicon Valley\u2019s next great leap may be built on videos of people folding laundry. Start-ups and entrepreneurs including&hellip;\n","protected":false},"author":2,"featured_media":589800,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[31],"tags":[74],"class_list":{"0":"post-589799","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-technology","8":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/589799","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/comments?post=589799"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/posts\/589799\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media\/589800"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/media?parent=589799"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/categories?post=589799"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/us\/wp-json\/wp\/v2\/tags?post=589799"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}