Research & Development World

  • R&D World Home
  • Topics
    • Aerospace
    • Automotive
    • Biotech
    • Careers
    • Chemistry
    • Environment
    • Energy
    • Life Science
    • Material Science
    • R&D Management
    • Physics
  • Technology
    • 3D Printing
    • A.I./Robotics
    • Software
    • Battery Technology
    • Controlled Environments
      • Cleanrooms
      • Graphene
      • Lasers
      • Regulations/Standards
      • Sensors
    • Imaging
    • Nanotechnology
    • Scientific Computing
      • Big Data
      • HPC/Supercomputing
      • Informatics
      • Security
    • Semiconductors
  • R&D Market Pulse
  • R&D 100
    • Call for Nominations: The 2025 R&D 100 Awards
    • R&D 100 Awards Event
    • R&D 100 Submissions
    • Winner Archive
    • Explore the 2024 R&D 100 award winners and finalists
  • Resources
    • Research Reports
    • Digital Issues
    • R&D Index
    • Subscribe
    • Video
    • Webinars
  • Global Funding Forecast
  • Top Labs
  • Advertise
  • SUBSCRIBE

Robotic Perception Steps toward Future Humanoid Robots

By R&D Editors | October 27, 2009

Robotic Perception Steps toward Future Humanoid Robots 

POPEYE, the audiovisual robot head
POPEYE, the audiovisual robot head Courtesy of POP project

European researchers have developed technology that enables a robot to combine data from both sound and vision to create combined, purposeful perception. In the process, they have taken the field to a new level.

Currently, computer vision is good at recognizing objects in images and videos and has been successfully employed in several specialized industrial applications, such as quality control during microchip fabrication. However, robotic perception is much weaker in less-defined situations, like understanding and responding to human behavior and even conversations. Yet, it is precisely this sort of interaction which promises the most compelling applications for future humanoid technology, where people-like robots can act as guides, or mix with people, or use perception to infer appropriate actions. More importantly, these broad robotic applications will deliver insights into other disciplines, like cognition and neuroscience.

A truly perceptive robot, capable of acting independently and appropriately in complex situations remains a distant goal, but European researchers have brought it much closer with their Perception-on-Purpose (POP) project.

“The originality of our project was our attempt to integrate two different sensory modalities, namely sound and vision,” explains Radu Horaud, POP’s coordinator. “This was very difficult to do, because you are integrating two completely different physical phenomena,” he adds.

Vision works from the reflection of light waves from an object, and it allows the observer to infer certain properties, like size, shape, density and texture. But, with sound you are interested in locating the direction of the source, and trying to identify the type of sound it is.

On its own, sound is difficult to pinpoint, because it needs to be located in a 3-D space. Then there is the problem of background noise, such as an open window letting in sounds from next door. However, it turned out that integrating two different senses helped the researchers in their bid to locate and tune into relevant sounds.

“It is not that easy to decide what is foreground and what is background using sound alone but, by combining the two modalities — sound and vision — it becomes much easier,” reveals Horaud. “If you are able to locate 10 sound sources in 10 different directions, but if in one of these directions you see a face, then you can much more easily concentrate on that sound and throw out the other ones.”

This was one approach that the team took and, with the algorithms they developed, their robot, called Popeye, was able to identify the speaker with a fair degree of reliability.

“There is more work to be done on that aspect of the work, it is not completely robust yet,” warns Horaud.

Still, it was a very strong result, and what makes it even more impressive is that the team managed to integrate all the technology into a neat and compact robotic platform.

“Most often, sound research is conducted in specialized labs, with arrays of microphones and a very controlled acoustic environment. But, we integrated our two microphones and two cameras onto the head of our Popeye. The idea is to have an agent-centered cognitive system,” Horaud stresses.

The Popeye packs a lot of powerful technology into a small space and offers purposeful robotic perception. This is important, because Horaud argues persuasively that, in evolutionary terms, multi-sensory perception and cognition are linked.

By perceiving a hand-held object with their two eyes, for example, monkeys — and the first hominids after them — developed stereo vision and hence were able to learn many properties of an object from combined tactile and visual data. Over time, they developed new skills, including building tools, from this information.

Horaud feels, too, that some modern uses of artificial intelligence (AI), like chess applications, are limited because they do not learn from their environment. They are programmed with abstract data — say, chess moves — and they process that.

“They cannot infer predicates from natural images; they cannot draw abstract information from physical observations,” he stresses.

For now, POP has achieved many of its aims and developed very promising approaches. Commercial applications for this type of technology are not out of the question, and the researchers also hope to continue their work in a further project. That project would look at extending some of POP’s results into a functioning humanoid robot. In the meantime, POP’s work means that the purposefully perceptive robot has become a not-so-distant future technology.

The POP project received funding from the Sixth Framework Programme for research.

Courtesy of ICT Results 

Related Articles Read More >

Why IBM predicts quantum advantage within two years
Aardvark AI forecasts rival supercomputer simulations while using over 99.9% less compute
This week in AI research: Latest Insilico Medicine drug enters the clinic, a $0.55/M token model R1 rivals OpenAI’s $60 flagship, and more
How the startup ALAFIA Supercomputers is deploying on-prem AI for medical research and clinical care
rd newsletter
EXPAND YOUR KNOWLEDGE AND STAY CONNECTED
Get the latest info on technologies, trends, and strategies in Research & Development.
RD 25 Power Index

R&D World Digital Issues

Fall 2024 issue

Browse the most current issue of R&D World and back issues in an easy to use high quality format. Clip, share and download with the leading R&D magazine today.

Research & Development World
  • Subscribe to R&D World Magazine
  • Enews Sign Up
  • Contact Us
  • About Us
  • Drug Discovery & Development
  • Pharmaceutical Processing
  • Global Funding Forecast

Copyright © 2025 WTWH Media LLC. All Rights Reserved. The material on this site may not be reproduced, distributed, transmitted, cached or otherwise used, except with the prior written permission of WTWH Media
Privacy Policy | Advertising | About Us

Search R&D World

  • R&D World Home
  • Topics
    • Aerospace
    • Automotive
    • Biotech
    • Careers
    • Chemistry
    • Environment
    • Energy
    • Life Science
    • Material Science
    • R&D Management
    • Physics
  • Technology
    • 3D Printing
    • A.I./Robotics
    • Software
    • Battery Technology
    • Controlled Environments
      • Cleanrooms
      • Graphene
      • Lasers
      • Regulations/Standards
      • Sensors
    • Imaging
    • Nanotechnology
    • Scientific Computing
      • Big Data
      • HPC/Supercomputing
      • Informatics
      • Security
    • Semiconductors
  • R&D Market Pulse
  • R&D 100
    • Call for Nominations: The 2025 R&D 100 Awards
    • R&D 100 Awards Event
    • R&D 100 Submissions
    • Winner Archive
    • Explore the 2024 R&D 100 award winners and finalists
  • Resources
    • Research Reports
    • Digital Issues
    • R&D Index
    • Subscribe
    • Video
    • Webinars
  • Global Funding Forecast
  • Top Labs
  • Advertise
  • SUBSCRIBE