Date(s) - 11/06/2020
10:00 am - 12:00 pm
|Speakers: Prof. Wenzhen Yuan|
With the rapid progress in robotics, people expect robots to be able to accomplish a wide variety of tasks in the real world, such as working in factories, performing household chores, and caring for elderly. However, it is still very difficult for robots to act in the physical world. A major challenge lies in the lack of adequate tactile sensing. Progress requires advances in the sensing hardware, but also requires advances in the software that can exploit the tactile signals generated when the robot touches an object. The sensor we use is a vision-based tactile sensor called GelSight, which measures the geometry and traction field of the contact surface. For interpreting the high-resolution tactile signal, we utilize both traditional statistical models and deep neural networks.
I will describe research on two kinds of tasks: exploration and manipulation. For exploration, I use active touch to estimate the physical properties of the objects. The work has included learning the basic properties (e.g., hardness), of artificial objects, as well as estimating the general properties of natural objects via autonomous tactile exploration. For manipulation, I study the robot’s ability to detect slip or incipient slip with tactile sensing during grasping. My research helpsrobots to better understand and flexibly interact with the physical world.
About the speaker:
Wenzhen Yuan is an assistant professor in the Robotics Insititute (RI) at Carnegie Mellon University. She received her PhD and Master degree fromMIT, surpervised by Prof. Edward Adelson and Dr. Mandayam Srinivasan. She received her BE degree from Tsinghua University. She also worked as a postdoctoral researcher at Stanford University. Her research interest is in tactile sensing, robotic perception and manipulation, and soft robotics.