fresh bytes
Subscribe Now

Robots learning to cook by watching YouTube videos

MSEEresearchers-1422902753519.jpg

The research we’re talking about here is from a paper titled, “Robot Learning Manipulation Action Plans by ‘Watching’ Unconstrained Videos
from the World Wide Web.” The paper is really about visual processing: watching a human interacting with objects in a video, and then figuring out what that human is doing and how they’re doing it, with a final step of replicating those actions using the manipulation capabilities of a robot (Baxter, in this case).

The University of Michigan has a dataset called YouCook, which consists of 88 open-source third-person YouTube cooking videos. Each video was given a set of unconstrained natural language descriptions by humans, and each video also has frame-by-frame object and action annotations. Using these data, the UMD researchers developed two convolutional neural networks: one to recognize and classify the objects in the videos, and the other to recognize and classify the grasps that the human is using.
via IEEE Spectrum

Continue reading 

Image: University of Maryland

Leave a Reply

featured blogs
Feb 6, 2026
In which we meet a super-sized Arduino Uno that is making me drool with desire....

featured chalk talk

Simplifying Position Control with Advanced Stepper Motor Driver
In this episode of Chalk Talk, Jiri Keprda from STMicroelectronics and Amelia Dalton explore the benefits of the powerSTEP01 is a system-in-package from STMicroelectronics. They also examine how this solution can streamline overall position control architecture, the high level commands included in this solution and the variety of advanced diagnostics included in the powerSTEP01 system-in-package.
Jan 21, 2025
31,019 views