{"id":1919,"date":"2026-03-06T09:30:00","date_gmt":"2026-03-06T14:30:00","guid":{"rendered":"https:\/\/techy44.okdiario.com\/en\/?p=1919"},"modified":"2026-03-06T10:32:57","modified_gmt":"2026-03-06T15:32:57","slug":"they-programmed-a-robot-to-watch-youtube-for-hours-and-what-it-learned-wasnt-programmed-into-it-its-a-bit-dizzying","status":"publish","type":"post","link":"https:\/\/techy44.okdiario.com\/en\/they-programmed-a-robot-to-watch-youtube-for-hours-and-what-it-learned-wasnt-programmed-into-it-its-a-bit-dizzying\/1919\/","title":{"rendered":"They programmed a robot to watch YouTube for hours\u2026 and what it learned wasn&#8217;t programmed into it (it&#8217;s a bit dizzying)"},"content":{"rendered":"\n<p>A soft faced humanoid robot has learned to move its lips in sync with speech and song simply by watching people talk on video and studying its own reflection in a mirror. <\/p>\n\n\n\n<p>The machine, called EMO, was built at Columbia Engineering and is at the center of a new study showing thatrobots can pick up complex speech related gestures through observation instead of hand written code. <\/p>\n\n\n\n<p>The work, which appears in the journal <a href=\"https:\/\/techy44.okdiario.com\/en\/a-b-2-spirit-joined-the-fighter-jets-of-carrier-air-wing-11-off-the-coast-of-california-and-the-novelty-was-the-aim-174b-gunslinger-on-the-super-hornets-designed-to-fire-at-long-ra\/25\/\">Science <\/a>Robotics, points to a future where robot conversations feel far less stiff and cartoonish than they do today.&nbsp;<\/p>\n\n\n\n<div class=\"gb-element-a00da4e5\">\n<div><div class=\"gb-looper-46613eed\">\n<div class=\"gb-loop-item gb-loop-item-a8390598 post-3451 post type-post status-publish format-standard has-post-thumbnail hentry category-business resize-featured-image\">\n<h4 class=\"gb-text gb-text-24a51617\">Also Read: <a href=\"https:\/\/techy44.okdiario.com\/en\/luxurys-worst-moment-in-years-is-starting-to-hit-hermes-gucci-and-lvmh-and-the-real-shock-is-that-even-the-safest-names-look-exposed\/3451\/\">Luxury\u2019s worst moment in years is starting to hit Herm\u00e8s, Gucci, and LVMH, and the real shock is that even the safest names look exposed<\/a><\/h4>\n<\/div>\n<\/div><\/div>\n<\/div>\n\n\n\n<p>The work, which appears in the journal Science Robotics, points to a future where robot conversations feel far less stiff and cartoonish than they do today. The work, which appears in the journal Science Robotics, points to a future where robot conversations feel far less stiff and cartoonish than they do today. <\/p>\n\n\n\n<p>The work, which appears in the journal Science Robotics, points to a future where robot conversations feel far less stiff and cartoonish than they do today.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Why faces and lips matter in conversation<\/h2>\n\n\n\n<p>If you have ever found yourself staring at someone\u2019s mouth while they speak, you are not alone. Eye tracking studies suggest humans devote a notable share of their attention to lips and lower face during conversation, which is one reason clumsy mouth motion makes many robots feel unsettling. EMO tries to solve that. <\/p>\n\n\n\n<p>Its silicone face is driven by 26 tiny motors that can pull and push the lips with fine control, more like human muscle than the rigid jaws seen on many social robots.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">How EMO trained itself using a mirror and YouTube<\/h2>\n\n\n\n<p>Training started with a kind of robotic mirror play. Engineers sat EMO in front of a reflective surface and let it fire off thousands of random expressions while a vision to action model learned how different motor patterns produced different mouth shapes. <\/p>\n\n\n\n<div class=\"gb-element-a1fe35a7\">\n<div><div class=\"gb-looper-814e445b\">\n<div class=\"gb-loop-item gb-loop-item-2007f8f1 post-1901 post type-post status-publish format-standard has-post-thumbnail hentry category-business resize-featured-image\">\n<h4 class=\"gb-text gb-text-0dd60053\">Also Read: <a href=\"https:\/\/techy44.okdiario.com\/en\/china-builds-a-high-speed-railway-line-in-9-hours-with-1500-workers-a-7-hour-journey-is-reduced-to-90-minutes\/1901\/\">China builds a high-speed railway line in 9 hours with 1,500 workers: a 7-hour journey is reduced to 90 minutes<\/a><\/h4>\n<\/div>\n<\/div><\/div>\n<\/div>\n\n\n\n<p>Once the system understood its own face, the team fed it hours of talking and singing clips on YouTube. By matching the sounds it heard with the lip positions it saw, the robot gradually learned to turn raw audio into the right sequence of facial movements across ten different languages.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">What people noticed in tests<\/h2>\n\n\n\n<p>To find out whether people actually bought the effect, the researchers showed videos of EMO speaking to more than one thousand volunteers. <\/p>\n\n\n\n<p>Viewers compared three different control methods against a reference of ideal lip motion and chose the new vision to action approach in roughly sixty two percent of trials, far ahead of the simpler baselines that only tracked loudness or copied past examples. <\/p>\n\n\n\n<figure class=\"wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio\"><div class=\"wp-block-embed__wrapper\">\n<iframe loading=\"lazy\" title=\"A Robot Learns to Lip Sync\" width=\"1200\" height=\"675\" src=\"https:\/\/www.youtube.com\/embed\/nhFU5KHA2fw?start=4&#038;feature=oembed\" frameborder=\"0\" allow=\"accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share\" referrerpolicy=\"strict-origin-when-cross-origin\" allowfullscreen><\/iframe>\n<\/div><figcaption class=\"wp-element-caption\">YouTube: <em>@ColumbiaSEAS<\/em>.<\/figcaption><\/figure>\n\n\n\n<p>Hard consonants such as B and sounds that require lip puckering still trip the system up, but the team expects performance to improve as EMO keeps \u201clistening\u201d and practicing.<\/p>\n\n\n\n<h2 class=\"wp-block-heading\">Why this could change human robot interaction<\/h2>\n\n\n\n<p>For the most part, the bigger story is what happens when this kind of realistic face is paired with conversational artificial intelligence. <\/p>\n\n\n\n<p>Lead author Yuhang Hu notes that combining fluent lip syncing with modern dialogue models could make exchanges with robots feel more like talking to another person than to a machine, especially in settings such as classrooms, hospitals, or elder care homes where empathy and trust matter.<\/p>\n\n\n\n<div class=\"gb-element-ea557750\">\n<div><div class=\"gb-looper-74e6aed1\">\n<div class=\"gb-loop-item gb-loop-item-86128e21 post-1895 post type-post status-publish format-standard has-post-thumbnail hentry category-military-defense resize-featured-image\">\n<h4 class=\"gb-text gb-text-9d7a26a3\">Also Read: <a href=\"https:\/\/techy44.okdiario.com\/en\/europe-accelerates-its-military-technology-in-response-to-russias-hypersonic-missile-the-terrifying-gap-and-the-race-to-close-it\/1895\/\">Europe accelerates its military technology in response to Russia&#8217;s hypersonic missile: the terrifying gap and the race to close it<\/a><\/h4>\n<\/div>\n<\/div><\/div>\n<\/div>\n\n\n\n<p>That possibility cuts both ways. Study supervisor Hod Lipson has warned that robots which smile and speak convincingly will be powerful tools and should be developed slowly and carefully so they help people without misleading them. <\/p>\n\n\n\n<p>If billions of humanoid machines are coming, as some economists suggest, then teaching them to \u201cuse their face\u201d responsibly may matter as much as teaching them to walk.<\/p>\n\n\n\n<p>The study was published on <em><a href=\"https:\/\/www.science.org\/doi\/10.1126\/scirobotics.adx3017\" target=\"_blank\" rel=\"noopener\">Science Robotics<\/a><\/em>.<\/p>\n\n\n\n<p><\/p>\n","protected":false},"excerpt":{"rendered":"<p>A soft faced humanoid robot has learned to move its lips in sync with speech and song simply by watching &#8230; <\/p>\n<p class=\"read-more-container\"><a title=\"They programmed a robot to watch YouTube for hours\u2026 and what it learned wasn&#8217;t programmed into it (it&#8217;s a bit dizzying)\" class=\"read-more button\" href=\"https:\/\/techy44.okdiario.com\/en\/they-programmed-a-robot-to-watch-youtube-for-hours-and-what-it-learned-wasnt-programmed-into-it-its-a-bit-dizzying\/1919\/#more-1919\" aria-label=\"Read more about They programmed a robot to watch YouTube for hours\u2026 and what it learned wasn&#8217;t programmed into it (it&#8217;s a bit dizzying)\">Read more<\/a><\/p>\n","protected":false},"author":6,"featured_media":1924,"comment_status":"open","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[10],"tags":[],"class_list":["post-1919","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-tech","resize-featured-image"],"_links":{"self":[{"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/posts\/1919","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/users\/6"}],"replies":[{"embeddable":true,"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/comments?post=1919"}],"version-history":[{"count":3,"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/posts\/1919\/revisions"}],"predecessor-version":[{"id":1955,"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/posts\/1919\/revisions\/1955"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/media\/1924"}],"wp:attachment":[{"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/media?parent=1919"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/categories?post=1919"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/techy44.okdiario.com\/en\/wp-json\/wp\/v2\/tags?post=1919"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}