by Staff Writers
Providence RI (SPX) May 18, 2012
On April 12, 2011, nearly fifteen years after she became paralyzed and unable to speak, a woman controlled a robotic arm by thinking about moving her arm and hand to lift a bottle of coffee to her mouth and take a drink.
That achievement is one of the advances in brain-computer interfaces restorative neurotechnology and assistive robot technology described in the journal Nature by the BrainGate2 collaboration of researchers at the Department of Veterans Affairs, Brown University, Massachusetts General Hospital, Harvard Medical School., and the German Aerospace Center (DLR).
A 58-year-old woman, "S3," and a 66-year-old man, "T2," participated in the study. They had each been paralyzed by a brainstem stroke years earlier which left them with no functional control of their limbs.
In the research, the participants used neural activity to directly control two different robotic arms, one developed by the DLR Institute of Robotics and Mechatronics and the other by DEKA Research and Development Corp., to perform reaching and grasping tasks across a broad three-dimensional space.
The BrainGate2 (www.braingate2.org) pilot clinical trial employs the investigational BrainGate system initially developed at Brown University, in which a baby aspirin-sized device with a grid of 96 tiny electrodes is implanted in the motor cortex-a part of the brain that is involved in voluntary movement. The electrodes are close enough to individual neurons to record the neural activity associated with intended movement.
An external computer translates the pattern of impulses across a population of neurons into commands to operate assistive devices, such as the DLR and DEKA robot arms used in the study now reported in Nature.
BrainGate participants have previously demonstrated neurally based two-dimensional point-and-click control of a cursor on a computer screen and rudimentary control of simple robotic devices.
The study represents the first demonstration and the first peer-reviewed report of people with tetraplegia using brain signals to control a robotic arm in three-dimensional space to complete a task usually performed by their arm. Specifically, S3 and T2 controlled the arms to reach for and grasp foam targets that were placed in front of them using flexible supports.
In addition, S3 used the DLR robot to pick up a bottle of coffee, bring it to her mouth, issue a command to tip it, drink through a straw, and return the bottle to the table. Her BrainGate-enabled, robotic-arm control during the drinking task required a combination of two-dimensional movements across a table top plus a "grasp" command to either grasp and lift or tilt the robotic hand.
"Our goal in this research is to develop technology that will restore independence and mobility for people with paralysis or limb loss," said lead author Dr. Leigh Hochberg, a neuroengineer and critical care neurologist who holds appointments at the Department of Veterans Affairs, Brown University, Massachusetts General Hospital and Harvard. He is the Sponsor-Investigator for the BrainGate2 pilot clinical trial.
"We have much more work to do, but the encouraging progress of this research is demonstrated not only in the reach-and-grasp data, but even more so in S3's smile when she served herself coffee of her own volition for the first time in almost 15 years."
Partial funding for this work comes from the VA, which is committed to improving the lives of injured Veterans. "VA is honored to have played a role in this exciting and promising area of research," said VA Secretary Eric Shinseki.
"Today's announcement represents a great step forward toward improving the quality of life for Veterans and others who have either lost limbs or are paralyzed."
Hochberg adds that even after nearly 15 years, a part of the brain essentially "disconnected" from its original target by a brainstem stroke was still able to direct the complex, multidimensional movement of an external arm - in this case, a robotic limb.
The researchers also noted that S3 was able to perform the tasks more than five years after the investigational BrainGate electrode array was implanted. This sets a new benchmark for how long implanted brain-computer interface electrodes have remained viable and provided useful command signals.
John Donoghue, the VA and Brown neuroscientist who pioneered BrainGate more than a decade ago and who is co-senior author of the study, said the paper shows how far the field of brain-computer interfaces has come since the first demonstrations of computer control with BrainGate.
"This paper reports an important advance by rigorously demonstrating in more than one participant that precise three-dimensional neural control of robot arms is not only possible, but also repeatable," said Donoghue, who directs the Brown Institute for Brain Science.
"We've moved significantly closer to returning everyday functions, like serving yourself a sip of coffee, usually performed effortlessly by the arm and hand, for people who are unable to move their own limbs. We are also encouraged to see useful control more than five years after implant of the BrainGate array in one of our participants. This work is a critical step toward realizing the long-term goal of creating a neurotechnology that will restore movement, control and independence to people with paralysis or limb loss."
In the research, the robots acted as a substitute for each participant's paralyzed arm. The robotic arms responded to the participants' intent to move as they imagined reaching for each foam target. The robot hand grasped the target when the participants imagined a hand squeeze. Because the diameter of the targets was more than half the width of the robot hand openings, the task required the participants to exert precise control. (Movies of these actions are also available on the Nature website.)
In 158 trials over four days, S3 was able to touch the target within an allotted time in 48.8 percent of the cases using the DLR robotic arm and hand and 69.2 percent of the cases with the DEKA arm and hand, which has the wider grasp. In 45 trials using the DEKA arm, T2 touched the target 95.6 percent of the time. Of the successful touches, S3 grasped the target 43.6 percent of the time with the DLR arm and 66.7 percent of the time with the DEKA arm. T2's grasp succeeded 62.2 percent of the time.
T2 performed the session in this study on his fourth day of interacting with the arm; the prior three sessions were focused on system development. Using his eyes to indicate each letter, he later described his control of the arm: "I just imagined moving my own arm and the [DEKA] arm moved where I wanted it to go."
The study used two advanced robotic arms: the DLR Light-Weight Robot III with DLR five-fingered hand and the DEKA Arm System. The DLR LWR-III, which is designed to assist in recreating actions like the human arm and hand and to interact with human users, could be valuable as an assistive robotic device for people with various disabilities.
Patrick van der Smagt, head of bionics and assistive robotics at DLR, director of biomimetic robotics and machine learning labs at DLR and the Technische Universitat Munchen, and a co-senior author on the paper said: "This is what we were hoping for with this arm. We wanted to create an arm that could be used intuitively by varying forms of control. The arm is already in use by numerous research labs around the world who use its unique interaction and safety capabilities. This is a compelling demonstration of the potential utility of the arm by a person with paralysis."
DEKA Research and Development developed the DEKA Arm System for amputees, through funding from the United States Defense Advanced Research Projects Agency (DARPA).
Dean Kamen, founder of DEKA said, "One of our dreams for the Luke Arm [as the DEKA Arm System is known informally] since its inception has been to provide a limb that could be operated not only by external sensors, but also by more directly thought-driven control. We're pleased about these results and for the continued research being done by the group at the VA, Brown and MGH."
The research is aimed at learning how the DEKA arm might be controlled directly from the brain, potentially allowing amputees to more naturally control this prosthetic limb.
Over the past two years, VA has been conducting an optimization study of the DEKA prosthetic arm at several sites, with the cooperation of Veterans and active duty service members who have lost an arm. Feedback from the study is helping DEKA engineers to refine the artificial arm's design and function.
"Brain-computer interfaces, such as BrainGate, have the potential to provide an unprecedented level of functional control over prosthetic arms of the future," said Joel Kupersmith, MD, VA Chief Research and Development Officer. "This innovation is an example of federal collaboration at its finest."
Story Landis, director of the National Institute of Neurological Disorders and Stroke, which funded the work in part, noted: "This technology was made possible by decades of investment and research into how the brain controls movement. It's been thrilling to see the technology evolve from studies of basic neurophysiology and move into clinical trials, where it is showing significant promise for people with brain injuries and disorders."
The BrainGate2 study continues to enroll participants to take part in this research and recently added Stanford University as a member of the collaboration and a clinical trial site.
In addition to Hochberg, Donoghue, and van der Smagt, other authors on the paper are Daniel Bacher, Beata Jarosiewicz, Nicolas Masse, John Simeral, Joern Vogel, Sami Haddadin, Jie Liu, and Sydney Cash. Additional comments can be found here.
All about the robots on Earth and beyond!
Comment on this article via your Facebook, Yahoo, AOL, Hotmail login.
Japan firm unveils gesture controlling device
Tokyo (AFP) May 16, 2012
Japanese technology titan NEC has unveiled a gadget that allows users to control their TV, mobile phone or tablet computer using a virtual input device. The company said a camera that recognises three-dimensional shapes and their movements will combine with a projector that could show the image of an input device - such as a keyboard - on almost any surface. Users can then gesture at t ... read more
|The content herein, unless otherwise known to be public domain, are Copyright 1995-2012 - Space Media Network. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA Portal Reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. Advertising does not imply endorsement,agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. Privacy Statement|