Research & Development World

  • R&D World Home
  • Topics
    • Aerospace
    • Automotive
    • Biotech
    • Careers
    • Chemistry
    • Environment
    • Energy
    • Life Science
    • Material Science
    • R&D Management
    • Physics
  • Technology
    • 3D Printing
    • A.I./Robotics
    • Software
    • Battery Technology
    • Controlled Environments
      • Cleanrooms
      • Graphene
      • Lasers
      • Regulations/Standards
      • Sensors
    • Imaging
    • Nanotechnology
    • Scientific Computing
      • Big Data
      • HPC/Supercomputing
      • Informatics
      • Security
    • Semiconductors
  • R&D Market Pulse
  • R&D 100
    • Call for Nominations: The 2025 R&D 100 Awards
    • R&D 100 Awards Event
    • R&D 100 Submissions
    • Winner Archive
    • Explore the 2024 R&D 100 award winners and finalists
  • Resources
    • Research Reports
    • Digital Issues
    • R&D Index
    • Subscribe
    • Video
    • Webinars
  • Global Funding Forecast
  • Top Labs
  • Advertise
  • SUBSCRIBE

Making Robots More Human

By R&D Editors | April 30, 2015

From joy to sadness, facial expressions could soon be decipherable to robots.Most people are naturally adept at reading facial expressions — from smiling and frowning to brow-furrowing and eye-rolling — to tell what others are feeling. Now, scientists have developed ultra-sensitive, wearable sensors that can do the same thing. Their technology, reported in the journal ACS Nano, could help robot developers make their machines more human.

Nae-Eung Lee and colleagues note that one way to make interactions between people and robots more intuitive would be to endow machines with the ability to read their users’ emotions and respond with a computer version of empathy. Most current efforts toward this goal analyze a person’s feelings using visual sensors that can tell a smile from a frown, for example. But these systems are expensive, highly complex and don’t pick up on subtle eye movements, which are important in human expression.

Lee’s team wanted to make simple, low-cost sensors to detect facial movements, including slight changes in gaze. The researchers created a stretchable and transparent sensor by layering a carbon nanotube film on two different kinds of electrically conductive elastomers. They found it could tell whether subjects were laughing or crying and where they were looking.

In addition to applications in robotics, the sensors could be used to monitor heartbeats, breathing, dysphagia (difficulty swallowing) and other health-related cues.

The authors acknowledge funding from the National Research Foundation of Korea.

Citation: “Stretchable, Transparent, Ultrasensitive, and Patchable Strain Sensor for Human-Machine Interfaces Comprising a Nanohybrid of Carbon Nanotubes and Conductive Elastomers”

  • Download full-text article

Related Articles Read More >

Why IBM predicts quantum advantage within two years
Aardvark AI forecasts rival supercomputer simulations while using over 99.9% less compute
This week in AI research: Latest Insilico Medicine drug enters the clinic, a $0.55/M token model R1 rivals OpenAI’s $60 flagship, and more
How the startup ALAFIA Supercomputers is deploying on-prem AI for medical research and clinical care
rd newsletter
EXPAND YOUR KNOWLEDGE AND STAY CONNECTED
Get the latest info on technologies, trends, and strategies in Research & Development.
RD 25 Power Index

R&D World Digital Issues

Fall 2024 issue

Browse the most current issue of R&D World and back issues in an easy to use high quality format. Clip, share and download with the leading R&D magazine today.

Research & Development World
  • Subscribe to R&D World Magazine
  • Enews Sign Up
  • Contact Us
  • About Us
  • Drug Discovery & Development
  • Pharmaceutical Processing
  • Global Funding Forecast

Copyright © 2025 WTWH Media LLC. All Rights Reserved. The material on this site may not be reproduced, distributed, transmitted, cached or otherwise used, except with the prior written permission of WTWH Media
Privacy Policy | Advertising | About Us

Search R&D World

  • R&D World Home
  • Topics
    • Aerospace
    • Automotive
    • Biotech
    • Careers
    • Chemistry
    • Environment
    • Energy
    • Life Science
    • Material Science
    • R&D Management
    • Physics
  • Technology
    • 3D Printing
    • A.I./Robotics
    • Software
    • Battery Technology
    • Controlled Environments
      • Cleanrooms
      • Graphene
      • Lasers
      • Regulations/Standards
      • Sensors
    • Imaging
    • Nanotechnology
    • Scientific Computing
      • Big Data
      • HPC/Supercomputing
      • Informatics
      • Security
    • Semiconductors
  • R&D Market Pulse
  • R&D 100
    • Call for Nominations: The 2025 R&D 100 Awards
    • R&D 100 Awards Event
    • R&D 100 Submissions
    • Winner Archive
    • Explore the 2024 R&D 100 award winners and finalists
  • Resources
    • Research Reports
    • Digital Issues
    • R&D Index
    • Subscribe
    • Video
    • Webinars
  • Global Funding Forecast
  • Top Labs
  • Advertise
  • SUBSCRIBE