24/7 Space News
AEROSPACE
A step toward safe and reliable autopilots for flying
MIT researchers developed a machine-learning technique that can autonomously drive a car or fly a plane through a very difficult "stabilize-avoid" scenario, in which the vehicle must stabilize its trajectory to arrive at and stay within some goal region, while avoiding obstacles.
A step toward safe and reliable autopilots for flying
by Adam Zewe for MIT News
Boston MA (SPX) Jun 12, 2023

In the film "Top Gun: Maverick," Maverick, played by Tom Cruise, is charged with training young pilots to complete a seemingly impossible mission - to fly their jets deep into a rocky canyon, staying so low to the ground they cannot be detected by radar, then rapidly climb out of the canyon at an extreme angle, avoiding the rock walls. Spoiler alert: With Maverick's help, these human pilots accomplish their mission.

A machine, on the other hand, would struggle to complete the same pulse-pounding task. To an autonomous aircraft, for instance, the most straightforward path toward the target is in conflict with what the machine needs to do to avoid colliding with the canyon walls or staying undetected. Many existing AI methods aren't able to overcome this conflict, known as the stabilize-avoid problem, and would be unable to reach their goal safely.

MIT researchers have developed a new technique that can solve complex stabilize-avoid problems better than other methods. Their machine-learning approach matches or exceeds the safety of existing methods while providing a tenfold increase in stability, meaning the agent reaches and remains stable within its goal region.

In an experiment that would make Maverick proud, their technique effectively piloted a simulated jet aircraft through a narrow corridor without crashing into the ground.

"This has been a longstanding, challenging problem. A lot of people have looked at it but didn't know how to handle such high-dimensional and complex dynamics," says Chuchu Fan, the Wilson Assistant Professor of Aeronautics and Astronautics, a member of the Laboratory for Information and Decision Systems (LIDS), and senior author of a new paper on this technique.

Fan is joined by lead author Oswin So, a graduate student. The paper will be presented at the Robotics: Science and Systems conference.

The stabilize-avoid challenge
Many approaches tackle complex stabilize-avoid problems by simplifying the system so they can solve it with straightforward math, but the simplified results often don't hold up to real-world dynamics.

More effective techniques use reinforcement learning, a machine-learning method where an agent learns by trial-and-error with a reward for behavior that gets it closer to a goal. But there are really two goals here - remain stable and avoid obstacles - and finding the right balance is tedious.

The MIT researchers broke the problem down into two steps. First, they reframe the stabilize-avoid problem as a constrained optimization problem. In this setup, solving the optimization enables the agent to reach and stabilize to its goal, meaning it stays within a certain region. By applying constraints, they ensure the agent avoids obstacles, So explains.

Then for the second step, they reformulate that constrained optimization problem into a mathematical representation known as the epigraph form and solve it using a deep reinforcement learning algorithm. The epigraph form lets them bypass the difficulties other methods face when using reinforcement learning.

"But deep reinforcement learning isn't designed to solve the epigraph form of an optimization problem, so we couldn't just plug it into our problem. We had to derive the mathematical expressions that work for our system. Once we had those new derivations, we combined them with some existing engineering tricks used by other methods," So says.

No points for second place
To test their approach, they designed a number of control experiments with different initial conditions. For instance, in some simulations, the autonomous agent needs to reach and stay inside a goal region while making drastic maneuvers to avoid obstacles that are on a collision course with it.

When compared with several baselines, their approach was the only one that could stabilize all trajectories while maintaining safety. To push their method even further, they used it to fly a simulated jet aircraft in a scenario one might see in a "Top Gun" movie. The jet had to stabilize to a target near the ground while maintaining a very low altitude and staying within a narrow flight corridor.

This simulated jet model was open-sourced in 2018 and had been designed by flight control experts as a testing challenge. Could researchers create a scenario that their controller could not fly? But the model was so complicated it was difficult to work with, and it still couldn't handle complex scenarios, Fan says.

The MIT researchers' controller was able to prevent the jet from crashing or stalling while stabilizing to the goal far better than any of the baselines.

In the future, this technique could be a starting point for designing controllers for highly dynamic robots that must meet safety and stability requirements, like autonomous delivery drones. Or it could be implemented as part of larger system. Perhaps the algorithm is only activated when a car skids on a snowy road to help the driver safely navigate back to a stable trajectory.

Navigating extreme scenarios that a human wouldn't be able to handle is where their approach really shines, So adds.

"We believe that a goal we should strive for as a field is to give reinforcement learning the safety and stability guarantees that we will need to provide us with assurance when we deploy these controllers on mission-critical systems. We think this is a promising first step toward achieving that goal," he says.

Moving forward, the researchers want to enhance their technique so it is better able to take uncertainty into account when solving the optimization. They also want to investigate how well the algorithm works when deployed on hardware, since there will be mismatches between the dynamics of the model and those in the real world.

The work is funded, in part, by MIT Lincoln Laboratory under the Safety in Aerobatic Flight Regimes program.

Research Report:"Solving Stabilize-Avoid Optimal Control via Epigraph Form and Deep Reinforcement Learning"

Related Links
Reliable Autonomous Systems Lab
Aerospace News at SpaceMart.com

Subscribe Free To Our Daily Newsletters
Tweet

RELATED CONTENT
The following news reports may link to other Space Media Network websites.
AEROSPACE
Air force chief appointed chairman of Ethiopian Airlines
Addis Ababa (AFP) June 12, 2023
Ethiopia has appointed the head of the country's air force, Lieutenant General Yilma Merdassa, as chairman of national carrier Ethiopian Airlines, the company announced on Monday. While commander of Ethiopia's air force, Yilma served on the company's board of directors "for the last two and half years," before assuming his new role on June 8, the carrier said. He replaced Girma Wake, a 79-year-old veteran of the African airline industry, who joined the company in 1965 and served as its CEO from ... read more

AEROSPACE
Schools, museums, libraries can apply to receive artifacts from NASA

Catastrophic failure assessment of sealed cabin for ultra large manned spacecraft

Shenzhou-16 spaceship transports seeds for breeding experiments

Boeing's first crewed space launch delayed, again

AEROSPACE
China's parachute system makes controllable landing of rocket boosters

Arianespace's next Ariane 5 mission to support France and Germany's space ambitions

China launches rocket with record payload

Iran unveils homegrown defense shield-busting hypersonic missile

AEROSPACE
Up and Over - Curiosity Is Heading East: Sol 3857

How NASA gives a name to every spot it studies on Mars

Science and sampling attempts at the Onahu Outcrop

Time To Try a New Route: Sols 3853-3856

AEROSPACE
Tianzhou 5 reconnects with Tiangong space station

China questions whether there is a new moon race afoot

Three Chinese astronauts return safely to Earth

Scientific experimental samples brought back to Earth, delivered to scientists

AEROSPACE
CNES, E-Space complete next-generation low earth orbit constellation study

HawkEye 360's Cluster 7 begins operation in record time

York Space Systems acquires Emergent Space Technologies

How activity in outer space will affect regional inequalities in the future

AEROSPACE
Foldable phased-array transmitters for small satellites

goTenna's mesh network demonstrates Oahu connectivity for U.S. military

Discharge test for launcher antenna

D-Orbit launches 11th orbital transportation mission in years

AEROSPACE
Photosynthesis, key to life on Earth, starts with a single photon

Phosphate, a key building block of life, found on Saturn's moon Enceladus

Plate tectonics not required for the emergence of life

Elusive planets play "hide and seek" with CHEOPS

AEROSPACE
Colorful Kuiper Belt puzzle solved by UH researchers

Juice deployments complete: final form for Jupiter

First observation of a Polar Cyclone on Uranus

Research 'solves' mystery of Jupiter's stunning colour changes

Subscribe Free To Our Daily Newsletters




The content herein, unless otherwise known to be public domain, are Copyright 1995-2024 - Space Media Network. All websites are published in Australia and are solely subject to Australian law and governed by Fair Use principals for news reporting and research purposes. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA news reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. All articles labeled "by Staff Writers" include reports supplied to Space Media Network by industry news wires, PR agencies, corporate press officers and the like. Such articles are individually curated and edited by Space Media Network staff on the basis of the report's information value to our industry and professional readership. Advertising does not imply endorsement, agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. General Data Protection Regulation (GDPR) Statement Our advertisers use various cookies and the like to deliver the best ad banner available at one time. All network advertising suppliers have GDPR policies (Legitimate Interest) that conform with EU regulations for data collection. By using our websites you consent to cookie based advertising. If you do not agree with this then you must stop using the websites from May 25, 2018. Privacy Statement. Additional information can be found here at About Us.