Research & Development World

  • R&D World Home
  • Topics
    • Aerospace
    • Automotive
    • Biotech
    • Careers
    • Chemistry
    • Environment
    • Energy
    • Life Science
    • Material Science
    • R&D Management
    • Physics
  • Technology
    • 3D Printing
    • A.I./Robotics
    • Software
    • Battery Technology
    • Controlled Environments
      • Cleanrooms
      • Graphene
      • Lasers
      • Regulations/Standards
      • Sensors
    • Imaging
    • Nanotechnology
    • Scientific Computing
      • Big Data
      • HPC/Supercomputing
      • Informatics
      • Security
    • Semiconductors
  • R&D Market Pulse
  • R&D 100
    • Call for Nominations: The 2025 R&D 100 Awards
    • R&D 100 Awards Event
    • R&D 100 Submissions
    • Winner Archive
    • Explore the 2024 R&D 100 award winners and finalists
  • Resources
    • Research Reports
    • Digital Issues
    • R&D Index
    • Subscribe
    • Video
    • Webinars
  • Global Funding Forecast
  • Top Labs
  • Advertise
  • SUBSCRIBE

Interactive system detects touch and gestures on any surface

By R&D Editors | October 9, 2012

This composite image shows how fingers and hands are computed in a new touch-activated system that projects onto walls and other surfaces and allows people to interact with their environment and each other. Image: Purdue University

People can let their fingers—and hands—do the talking with a new touch-activated system that projects onto walls and other surfaces and allows users to interact with their environment and each other.

The system identifies the fingers of a person’s hand while touching any plain surface. It also recognizes hand posture and gestures, revealing individual users by their unique traits.

“Imagine having giant iPads everywhere, on any wall in your house or office, every kitchen counter, without using expensive technology,” said Niklas Elmqvist, an assistant professor of electrical and computer engineering at Purdue University. “You can use any surface, even a dumb physical surface like wood. You don’t need to install expensive LED displays and touch-sensitive screens.”

The new “extended multitouch” system allows more than one person to use a surface at the same time and also enables people to use both hands, distinguishing between the right and left hand.

Research indicates the system is 98% accurate in determining hand posture, which is critical to recognizing gestures and carrying out commands. The technology has many possible applications, said Karthik Ramani, Purdue’s Donald W. Feddersen Professor of Mechanical Engineering.

“Basically, it might be used for any interior surface to interact virtually with a computer,” he said. “You could use it for living environments, to turn appliances on, in a design studio to work on a concept or in a laboratory, where a student and instructor interact.”

Findings are detailed in a research paper being presented this week during the Association for Computing Machinery Symposium on User Interface Software and Technology (ACM UIST 2012) in Cambridge, Mass. The paper was written by doctoral students Sundar Murugappan and Vinayak, who uses only one name, Elmqvist and Ramani.

The system uses the Microsoft Kinect camera, which senses three-dimensional space.

“We project a computer screen on any surface, just a normal table covered with white paper,” Ramani said. “The camera sees where your hands are, which fingers you are pressing on the surface, tracks hand gestures and recognizes whether there is more than one person working at the same time.”

The Kinect camera senses depth, making it possible to see how far each 3-D pixel is from the camera. The researchers married the camera with a new computer model for the hand.

“So we can isolate different parts of a hand or finger to show how far they are from the surface,” Elmqvist said. “We can see which fingers are touching the surface. With this technology, you could potentially call up a menu by positioning your hand just above the surface.”

That camera coupled with the hand model allows the system to locate the center of each hand, which is necessary for determining gestures and distinguishing between left and right hands.
Researchers explored possible applications, including one that allows the user to draw a sketch with a pen and then modify it with their hands.

“We can detect gestural interactions between more than one hand and more than one user,” Ramani said. “You could do precision things, like writing with a pen, with your dominant hand and more general things, such as selecting colors, using the non-dominant hand.”

Researchers tested the concept in two user studies, one with 14 volunteers and the other with nine. Findings from one study indicated display features should be no smaller than 18 millimeters, or about an inch, to be efficient.

“While new and more precise cameras will improve accuracy, we have established the necessary hand models and principles for the system,” Ramani said.

The other user study showed the system can effectively determine hand posture and whether the right or left hand is being used.

“We wanted to see how accurate we could be while figuring out different configurations, such as touching with all 10 fingers, which hand is being used and so on,” Elmqvist said.

Patents are pending on the concept.

The research has been supported by the National Science Foundation and the Donald W. Feddersen Chaired Professorship at Purdue School of Mechanical Engineering.

Extended Multitouch: Recovering Touch Posture and Differentiating Users using a Depth Camera

Source: Purdue University

Related Articles Read More >

First CRISPR-edited spider spins red fluorescent silk
KIST carbon nanotube supercapacitor holds capacity after 100,000 cycles
A new wave of metalworking lets semiconductor crystals bend and stretch
LLNL deposits quantum dots on corrugated IR chips in a single step
rd newsletter
EXPAND YOUR KNOWLEDGE AND STAY CONNECTED
Get the latest info on technologies, trends, and strategies in Research & Development.
RD 25 Power Index

R&D World Digital Issues

Fall 2024 issue

Browse the most current issue of R&D World and back issues in an easy to use high quality format. Clip, share and download with the leading R&D magazine today.

Research & Development World
  • Subscribe to R&D World Magazine
  • Enews Sign Up
  • Contact Us
  • About Us
  • Drug Discovery & Development
  • Pharmaceutical Processing
  • Global Funding Forecast

Copyright © 2025 WTWH Media LLC. All Rights Reserved. The material on this site may not be reproduced, distributed, transmitted, cached or otherwise used, except with the prior written permission of WTWH Media
Privacy Policy | Advertising | About Us

Search R&D World

  • R&D World Home
  • Topics
    • Aerospace
    • Automotive
    • Biotech
    • Careers
    • Chemistry
    • Environment
    • Energy
    • Life Science
    • Material Science
    • R&D Management
    • Physics
  • Technology
    • 3D Printing
    • A.I./Robotics
    • Software
    • Battery Technology
    • Controlled Environments
      • Cleanrooms
      • Graphene
      • Lasers
      • Regulations/Standards
      • Sensors
    • Imaging
    • Nanotechnology
    • Scientific Computing
      • Big Data
      • HPC/Supercomputing
      • Informatics
      • Security
    • Semiconductors
  • R&D Market Pulse
  • R&D 100
    • Call for Nominations: The 2025 R&D 100 Awards
    • R&D 100 Awards Event
    • R&D 100 Submissions
    • Winner Archive
    • Explore the 2024 R&D 100 award winners and finalists
  • Resources
    • Research Reports
    • Digital Issues
    • R&D Index
    • Subscribe
    • Video
    • Webinars
  • Global Funding Forecast
  • Top Labs
  • Advertise
  • SUBSCRIBE