News Report SMW Technology
June 26, 2023

Meta AI Develops an Algorithm That Enables Robots to Learn Tasks from YouTube Videos

In Brief

Researchers have developed a visual affordance model using internet videos of human behavior to train robots to perform complex tasks.

This approach bridges the gap between static datasets and real-world robot applications.

The researchers use large-scale human video datasets like Ego4D and Epic Kitchens to extract affordances, integrating computer vision techniques with robotic manipulation.

The Vision-Robotics Bridge (VRB) concept showcases the potential of this approach, enabling robots to learn from human videos and acquire the skills necessary for complex tasks.

Meta AI unveiled a new algorithm that enables robots to learn and replicate human actions by watching YouTube videos. In a recent paper entitled “Affordances from Human Videos as a Versatile Representation for Robotics,” the authors explore how videos of human interactions can be leveraged to train robots to perform complex tasks.

Meta AI Develops an Algorithm That Enables Robots to Learn Tasks from YouTube Videos
Credit: Metaverse Post (mpost.io)

This research aims to bridge the gap between static datasets and real-world robot applications. While previous models have shown success on static datasets, applying these models directly to robots has remained a challenge. The researchers propose training a visual affordance model using internet videos of human behavior could be a solution. This model estimates where and how a human is likely to interact in a scene, providing valuable information for robots.

The concept of “affordances” is central to this approach. Affordances refer to the potential actions or interactions an object or environment offers. By understanding affordances through human videos, the robot gains a versatile representation that enables it to perform various complex tasks. The researchers integrate their affordance model with four different robot learning paradigms: offline imitation learning, exploration, goal-conditioned learning, and action parameterization for reinforcement learning.

Recommended: Top 100+ Words Detectable by AI Detectors in 2023

To extract affordances, the researchers utilize large-scale human video datasets like Ego4D and Epic Kitchens. They employ off-the-shelf hand-object interaction detectors to identify the contact region and track the wrist’s trajectory after contact. However, an important challenge arises when the human is still present in the scene, causing a distribution shift. To address this, the researchers use available camera information to project the contact points and post-contact trajectory to a human-agnostic frame, which serves as input to their model.

Previously, robots were capable of mimicking actions, but their abilities were limited to replicating specific environments. With the latest algorithm, researchers have made significant progress in “generalizing” robot actions. Robots can now apply their acquired knowledge in new and unfamiliar environments. This achievement aligns with the vision of achieving Artificial General Intelligence (AGI) as advocated by AI researcher Jan LeCun.

Recommended: GPT-4 Solves MIT Exam Questions with 100% Accuracy? Not True, Researchers Say
The model takes the human-agnostic frame as input and produces two key outputs
The model takes the human-agnostic frame as input and produces two key outputs: a contact heatmap and wrist waypoints. The contact heatmap indicates the likely points of contact, while the wrist waypoints predict the trajectory after contact. These outputs can be directly used during inference time, leveraging sparse 3D information such as depth and robot kinematics. / Credit: robo-affordances.github.io

Meta AI is committed to advancing the field of computer vision and is planning to share its project’s code and dataset. This will enable other researchers and developers to further explore and build upon this technology. With increased access to the code and dataset, the development of self-learning robots capable of acquiring new skills from YouTube videos will continue to progress.

The model takes the human-agnostic frame as input and produces two key outputs
To evaluate the efficacy of their approach, the researchers conducted experiments across four real-world environments involving ten different tasks and two robotic platforms operating in the wild. The results demonstrated the seamless integration of computer vision techniques with robotic manipulation, showcasing the potential of their Vision-Robotics Bridge (VRB) concept. / Credit: robo-affordances.github.io

By leveraging the vast amount of online instructional videos, robots can become more versatile and adaptable in various environments.

Read more about AI:

Disclaimer

In line with the Trust Project guidelines, please note that the information provided on this page is not intended to be and should not be interpreted as legal, tax, investment, financial, or any other form of advice. It is important to only invest what you can afford to lose and to seek independent financial advice if you have any doubts. For further information, we suggest referring to the terms and conditions as well as the help and support pages provided by the issuer or advertiser. MetaversePost is committed to accurate, unbiased reporting, but market conditions are subject to change without notice.

About The Author

Damir is the team leader, product manager, and editor at Metaverse Post, covering topics such as AI/ML, AGI, LLMs, Metaverse, and Web3-related fields. His articles attract a massive audience of over a million users every month. He appears to be an expert with 10 years of experience in SEO and digital marketing. Damir has been mentioned in Mashable, Wired, Cointelegraph, The New Yorker, Inside.com, Entrepreneur, BeInCrypto, and other publications. He travels between the UAE, Turkey, Russia, and the CIS as a digital nomad. Damir earned a bachelor's degree in physics, which he believes has given him the critical thinking skills needed to be successful in the ever-changing landscape of the internet. 

More articles
Damir Yalalov
Damir Yalalov

Damir is the team leader, product manager, and editor at Metaverse Post, covering topics such as AI/ML, AGI, LLMs, Metaverse, and Web3-related fields. His articles attract a massive audience of over a million users every month. He appears to be an expert with 10 years of experience in SEO and digital marketing. Damir has been mentioned in Mashable, Wired, Cointelegraph, The New Yorker, Inside.com, Entrepreneur, BeInCrypto, and other publications. He travels between the UAE, Turkey, Russia, and the CIS as a digital nomad. Damir earned a bachelor's degree in physics, which he believes has given him the critical thinking skills needed to be successful in the ever-changing landscape of the internet. 

Hot Stories
Join Our Newsletter.
Latest News

Sentencing Day Arrives: CZ’s Fate Hangs in Balance as US Court Considers DOJ’s Plea

Changpeng Zhao is poised to face sentencing in a U.S. court in Seattle today.

Know More

Samourai Wallet Founders Accused of Facilitating $2B in Darknet Deals

The apprehension of the Samourai Wallet founders represents a notable setback for the industry, underscoring the persistent ...

Know More
Join Our Innovative Tech Community
Read More
Read more
Galxe Partners With Jambo To  Expand Global Accessibility To Web3
Business News Report Technology
Galxe Partners With Jambo To Expand Global Accessibility To Web3
May 2, 2024
Google’s Med-Gemini Poised to Give a Head Start to GPT-4 With Its Superior Performance in Healthcare
AI Wiki News Software Technology
Google’s Med-Gemini Poised to Give a Head Start to GPT-4 With Its Superior Performance in Healthcare
May 2, 2024
Hong Kong Legislative Council Member Wu Jiezhuang Signals Civil Suit Against JPEX Crypto Exchange
Business News Report Technology
Hong Kong Legislative Council Member Wu Jiezhuang Signals Civil Suit Against JPEX Crypto Exchange
May 2, 2024
AltLayer Enters Second Phase Of Its Staking Initiative, Introduces reALT Token
Markets News Report Technology
AltLayer Enters Second Phase Of Its Staking Initiative, Introduces reALT Token
May 2, 2024