by Staff Writers
Munich, Germanay (SPX) Jun 01, 2011
Before a new vehicle rolls off the assembly lines, it first takes shape as a virtual model. In a cave - a room for the virtual representation of objects - the developers look at it from all sides, they sit in it, they examine and improve it. For example, are all the switches easy to reach?
The developers have so far used a joystick to interact with the computer which displays the virtual car model. In the future, they will be able to do so without such an aid - their hand alone is intended to be enough to provide the computer with the respective signals.
A multi-touch interface, which h was developed by Georg Hackenberg during his Master's thesis work at the Fraunhofer Institute for Applied Information Technology FIT, made this possible.
His work earned him first place in the Hugo Geiger Prizes. We are using a camera that, instead of providing color information, provides pixel for pixel the distance of how far this point is from the camera.
Basically this is achieved by means of a type of gray-scale image where the shade of gray represents the distance of the objects. The camera also provides three-dimensional information that the system evaluates with the help of special algorithms, explains Georg Hackenberg.
Hackenberg's main work consisted in developing the corresponding algorithms. They ensure that the system is first able to recognize a hand and then able to follow its movements.
The result: The 3D camera system processes gestures down to the movements of individual fingers and processes them in real time. Up to this point in time comparable processes with finger support could only detect how hands moved in the image level - they could not solve the depth information, in other words, how far the hand is from the camera system.
For this reason it was often difficult to answer with which object the hand was interacting. Is it activating the windshield wipers or is it turning on the radio?
Small movements of the hand, such as gripping, have so far been hardly possible to detect in real time - or only with great amounts of computing power. That is no problem for the new system.
Gesture commands are also interesting for computer games. A gesture recognition prototype already exists. The researchers want to improve weaknesses in the algorithm now and carry out initial application studies.
Hackenberg hopes that the system could be ready for series production within a year, from a technical viewpoint. In the medium term, the researchers hope to further develop it such that it can be used in mobile applications as well, which means that it will also find its way into laptops and cell phones
Comment on this article via your Facebook, Yahoo, AOL, Hotmail login.
Guide vests robotic navigation aids for the visually impaired
Los Angeles CA (SPX) May 27, 2011
For the visually impaired, navigating city streets or neighborhoods has constant challenges. And most such people still must rely on a very rudimentary technology - a simple cane - to help them make their way through a complex world. A group of University of Southern California engineering researchers is working to change that by developing a robot vision-based mobility aid for the visuall ... read more
|The content herein, unless otherwise known to be public domain, are Copyright 1995-2011 - Space Media Network. AFP and UPI Wire Stories are copyright Agence France-Presse and United Press International. ESA Portal Reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. Advertising does not imply endorsement,agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. Privacy Statement|