Categories: AI/ML News

Listening skills bring human-like touch to robots

Researchers give robots a sense of touch by ‘listening’ to vibrations, allowing them to identify materials, understand shapes and recognize objects just like human hands. The ability to interpret the world through acoustic vibrations emanating from an object — like shaking a cup to see how much soda is left or tapping on a desk to see if it’s made out of real wood — is something humans do without thinking. And it’s an ability that researchers are on the cusp of bringing to robots to augment their rapidly growing set of sensing abilities.
AI Generated Robotic Content

Share
Published by
AI Generated Robotic Content

Recent Posts

SamsungCam UltraReal – Qwen-Image LoRA

Hey everyone, Just dropped the first version of a LoRA I've been working on: SamsungCam…

8 hours ago

40 Best Early Amazon Prime Day Deals on WIRED-Tested Gear (2025)

Amazon Prime Day is back, starting on October 7, but we’ve already found good deals…

9 hours ago

These little robots literally walk on water

HydroSpread, a breakthrough fabrication method, lets scientists build ultrathin soft robots directly on water. These…

9 hours ago

VHS filters work great with AI footage (WAN 2.2 + NTSC-RS)

submitted by /u/mtrx3 [link] [comments]

1 day ago

Algorithm Showdown: Logistic Regression vs. Random Forest vs. XGBoost on Imbalanced Data

Imbalanced datasets are a common challenge in machine learning.

1 day ago

Unlock global AI inference scalability using new global cross-Region inference on Amazon Bedrock with Anthropic’s Claude Sonnet 4.5

Organizations are increasingly integrating generative AI capabilities into their applications to enhance customer experiences, streamline…

1 day ago