Imaginative and prescient-language-action fashions are the present cutting-edge in robotic manipulation. They nonetheless can not decide up a potato chip with out crushing it.
That’s the outcome revealed earlier this yr by the staff behind the Video Tactile Action Model (VTAM). On a potato chip pick-and-place activity — a activity that calls for high-fidelity pressure consciousness, the place imaginative and prescient alone can not distinguish a crushing grasp from a holding one — VTAM outperformed the π0.5 baseline by 80%. Throughout the broader contact-rich benchmark suite, VTAM held a 90% common success charge.¹
The chip is an adversarial instance, and that’s exactly why it’s the proper check. On the level of grasp, solely contact dynamics carry helpful indicators. Stress, vibration, and pressure/torque inform the coverage what is going on, correcting the visible estimation errors that vision-only fashions can not detect on their very own. A digicam, nonetheless excessive its decision, can not do this work.
Tactile shouldn’t be plug-and-play
Tactile sensors don’t enhance mannequin efficiency on their very own. Most studying pipelines at present are constructed round imaginative and prescient and language; the 2 modalities with the biggest datasets and essentially the most mature architectures behind them. When tactile indicators are appended to a vision-first pipeline with out intentional design, they have an inclination to get downweighted, drowned out, or misplaced in coaching. VTAM works as a result of the structure forces the mannequin to forecast imaginative and prescient and tactile dynamics collectively, so the tactile sign straight shapes the realized coverage slightly than getting absorbed into imaginative and prescient and language. Tactile knowledge solely delivers its worth when it’s intelligently included.
The sample is now constant throughout the literature
The chip is one finish of the spectrum, a case the place imaginative and prescient fails outright and tactile carries the sign alone. Most real-world duties sit additional alongside that spectrum, the place imaginative and prescient and tactile every contribute and the synergy between them is what drives coaching effectivity. The sample is now constant throughout the literature.
VTAM shouldn’t be alone. The ManiSkill-ViTac 2025 benchmark formalises tactile-augmented analysis throughout insertion, software use, and precision meeting duties. Impartial analysis on tactile sensor configurations and grasp studying efficiency² reveals the identical carry. Insurance policies that mix imaginative and prescient with tactile suggestions constantly outperform vision-only equivalents on contact-rich duties, and have a tendency to achieve the identical success threshold from fewer demonstrations.
Failure detection is the second prize
A tactile-conditioned coverage registers incipient slip as a vibration signature tens to a whole bunch of milliseconds earlier than the article truly strikes. That window is the distinction between re-grasping and a full restart — between 95% and 99% uptime on the identical line. Throughout a fleet, the operational case turns into exhausting to disregard.
Failure detection is one case of a bigger functionality: producing correct, high-resolution labels for what truly occurred through the grasp. A binary success/fail label collapses info that the coaching pipeline might use. Did the grasp succeed cleanly, or did it succeed with inside slippage that the controller recovered from? Did the article settle stably, or did it shift throughout transport? Tactile sensing can distinguish these circumstances, and embedded contact notion can label them on-device, turning each episode right into a extra informative coaching instance, not simply the failed ones.
Determine 1. VTAM combines a language mannequin, a predictive vision-tactile world mannequin, and a diffusion-based motion coverage. From simply 10 minutes of teleoperation per activity, it learns to foretell future actions, states, and forces — enabling contact-rich duties comparable to chip pick-and-place, dynamic wiping, and steady peeling. Supply: arXiv:2603.23481.
What this implies for builders
Tactile sensing has moved from helpful addition to defensible requirement for any staff aiming at production-grade contact-rich manipulation. The query is not whether or not to instrument. It’s whether or not to instrument now, or pay later in rebuilt datasets and recalibrated fashions.
VTAM put an actual quantity on the case and different latest work retains pointing in the identical path. The subsequent era of basis fashions will probably be constructed on knowledge that captures contact slightly than vision-only.
Able to take the following step?
Talk to our technical team about tactile integration on your manipulation pipeline and be taught extra about how Robotiq can enable your application.
¹ Video Tactile Motion Mannequin (VTAM), arXiv:2603.23481.
² Consultant findings embody Tactile Robotics: An Outlook (arXiv) and revealed work on the impression of tactile sensor configurations on grasp studying effectivity.

