Imagine a robot trying to pick up a delicate glass ornament or perform intricate surgery. These tasks require more than just mechanical grip—they demand the nuanced, skillful handling that humans perform effortlessly every day. This capability, known as dexterous manipulation, has long been one of the most challenging frontiers in robotics.
While modern robots have become remarkably adept at basic tasks like picking up objects and moving them from one place to another, they still struggle with the sophisticated handling that defines human dexterity. A significant part of this challenge stems from how robots perceive the world around them.

Most existing robotic systems rely heavily on visual sensors, such as cameras, to understand their environment. They can see an object, calculate its position, and plan a path to grab it. However, these systems often falter when it comes to what happens during and after contact—the moment when a robot's gripper actually touches and must manipulate an object.
The problem lies in the sensory gap. Traditional tactile sensors only provide feedback after physical contact has been made, leaving robots essentially blind during the critical moment of initial contact and afterwards. This limitation makes it remarkably difficult for robots to adjust their grip strength, reposition their hold, or respond to unexpected object movements.
Researchers at the National University of Singapore and RoboScience have now developed an innovative solution called FingerEye. This groundbreaking sensor represents a new approach to robotic perception, combining both visual and tactile information into a continuous stream that works before, during, and after contact with objects.
"Dexterous robotic manipulation requires comprehensive perception across all phases of interaction: pre-contact, contact initiation, and post-contact," explained Zhixuan Xu, Yuchen Li, and their colleagues in their research paper. "Such continuous feedback allows a robot to adapt its actions throughout interaction. However, many existing tactile sensors, such as GelSight and its variants, only provide feedback after contact is established, limiting a robot's ability to precisely initiate contact."
FingerEye: A new vision-tactile sensor
The FingerEye sensor represents a significant departure from conventional approaches to robotic sensing. Its design integrates two small colored cameras and a soft, flexible ring structure that work together to provide comprehensive perceptual information.
The dual camera system continuously captures images of nearby objects, employing a binocular approach similar to human depth perception. By estimating the distance between their two distinct viewpoints, the cameras can determine how far objects are from the sensor—a capability known as stereo depth estimation.
The ring structure surrounding the cameras serves a different but equally important purpose. When the sensor makes contact with an object, this soft ring deforms in response to applied forces and movements. By analyzing these deformations through marker-based pose estimation, the system can infer the exact forces and torques being applied during manipulation.
"FingerEye integrates binocular RGB cameras to provide close-range visual perception with implicit stereo depth," wrote the research team. "Upon contact, external forces and torques deform a compliant ring structure; these deformations are captured via marker-based pose estimation and serve as a proxy for contact wrench sensing."
One of the most notable aspects of the FingerEye design is its compact size and cost-effectiveness. Unlike many advanced robotic sensors that require expensive equipment or bulky installations, FingerEye can be integrated into existing robotic grippers without significant modifications. Most importantly, its unique architecture enables the collection of uninterrupted visual and tactile data throughout the entire manipulation process.
"This design enables a perception stream that smoothly transitions from pre-contact visual cues to post-contact tactile feedback," the researchers noted. Building on this capability, the team developed a vision-tactile imitation learning policy that combines signals from multiple FingerEye sensors to learn dexterous manipulation behaviors from limited real-world data.
In addition to the sensor itself, the researchers created a virtual replica, essentially a digital twin, of FingerEye. This digital model allows researchers to simulate and fine-tune robotic manipulation strategies in a virtual environment before testing them with physical robots, significantly accelerating the development process.
Initial results and real-world applications
The research team conducted extensive testing of their sensor both in simulations and real-world experiments. They integrated FingerEye onto robotic grippers and evaluated the system's performance across various manipulation tasks.
The results were highly encouraging. The sensor enabled robots to perceive objects continuously throughout the manipulation process—before contact, during contact, and after contact. This continuous perception led to noticeably more effective manipulation strategies, as robots could now adapt their actions in real-time based on immediate sensory feedback.
The team's work has been made available to the broader research community. Both the sensor's design specifications and the associated programming code are open-source and accessible through GitHub, allowing other researchers to build upon and learn from their work.
"By combining real demonstrations with visually augmented simulated observations for representation learning, the learned policies become more robust to object appearance variations," the authors explained. "Together, these design aspects enable dexterous manipulation across diverse object properties and interaction regimes, including coin standing, chip picking, letter retrieving, and syringe manipulation."
The implications of this technology extend far beyond the laboratory. FingerEye could pave the way for more adaptive robots capable of performing complex manual tasks in homes, hospitals, and industrial settings. From assisting elderly individuals with daily activities to performing delicate medical procedures, the applications are vast and varied.
As the technology continues to evolve, researchers anticipate that sensors like FingerEye will become increasingly sophisticated, enabling the next generation of robotic systems to handle tasks that currently require human intuition and skill. The bridge between vision and touch in robotics has finally been constructed, opening doors to capabilities that were previously impossible.
Source credit: TechXplore
Image credits:
- Image 1 - credit: TechXplore
- Left: FingerEye provides continuous vision-tactile perception across all phases of interaction. Before contact, binocular RGB cameras provide close-range visual cues and implicit stereo depth to guide fingertip positioning. Upon contact, external forces and torques deform a compliant ring structure; marker-based pose estimation converts these deformations into contact wrench signals for contact monitoring and control. Right: real-world evaluation tasks spanning diverse object properties and interaction regimes, including Chip Picking (rigid, delicate), Coin Standing (rigid, contact-sensitive), Letter Retrieving (deformable, thin-shell), and Syringe Manipulation (articulated, functional). Credit: Xu et al. - credit: TechXplore

Your Opinion is valid .