Just a sample of the Echomail archive
Cooperative anarchy at its finest, still active today. Darkrealms is the Zone 1 Hub.
|    EARTH    |    Uhh, that 3rd rock from the sun?    |    8,931 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 6,060 of 8,931    |
|    ScienceDaily to All    |
|    Rapid adaptation of deep learning teache    |
|    05 May 22 22:30:38    |
      MSGID: 1:317/3 6274a4ac       PID: hpt/lnx 1.9.0-cur 2019-01-08       TID: hpt/lnx 1.9.0-cur 2019-01-08        Rapid adaptation of deep learning teaches drones to survive any weather                      Date:        May 5, 2022        Source:        California Institute of Technology        Summary:        Neural-Fly technology could one day build the future of package        delivery drones and flying cars.                            FULL STORY       ==========================================================================       To be truly useful, drones -- that is, autonomous flying vehicles --       will need to learn to navigate real-world weather and wind conditions.                     ==========================================================================       Right now, drones are either flown under controlled conditions, with       no wind, or are operated by humans using remote controls. Drones have       been taught to fly in formation in the open skies, but those flights       are usually conducted under ideal conditions and circumstances.              However, for drones to autonomously perform necessary but quotidian tasks,       such as delivering packages or airlifting injured drivers from a traffic       accident, drones must be able to adapt to wind conditions in real time --       rolling with the punches, meteorologically speaking.              To face this challenge, a team of engineers from Caltech has developed       Neural- Fly, a deep-learning method that can help drones cope with       new and unknown wind conditions in real time just by updating a few       key parameters.              Neural-Fly is described in a study published on May 4 in Science       Robotics. The corresponding author is Soon-Jo Chung, Bren Professor of       Aerospace and Control and Dynamical Systems and Jet Propulsion Laboratory       Research Scientist. Caltech graduate students Michael O'Connell (MS '18)       and Guanya Shi are the co-first authors.              Neural-Fly was tested at Caltech's Center for Autonomous Systems       and Technologies (CAST) using its Real Weather Wind Tunnel, a custom       10-foot-by-10- foot array of more than 1,200 tiny computer-controlled fans       that allows engineers to simulate everything from a light gust to a gale.                            ==========================================================================       "The issue is that the direct and specific effect of various wind       conditions on aircraft dynamics, performance, and stability cannot       be accurately characterized as a simple mathematical model," Chung       says. "Rather than try to qualify and quantify each and every effect       of turbulent and unpredictable wind conditions we often experience       in air travel, we instead employ a combined approach of deep learning       and adaptive control that allows the aircraft to learn from previous       experiences and adapt to new conditions on the fly with stability and       robustness guarantees." O'Connell adds: "We have many different models       derived from fluid mechanics, but achieving the right model fidelity and       tuning that model for each vehicle, wind condition, and operating mode is       challenging. On the other hand, existing machine learning methods require       huge amounts of data to train yet do not match state-of-the-art flight       performance achieved using classical physics-based methods. Moreover,       adapting an entire deep neural network in real time is a huge, if not       currently impossible task." Neural-Fly, the researchers say, gets around       these challenges by using a so- called separation strategy, through which       only a few parameters of the neural network must be updated in real time.              "This is achieved with our new meta-learning algorithm, which pre-trains       the neural network so that only these key parameters need to be updated       to effectively capture the changing environment," Shi says.              After obtaining as little as 12 minutes of flying data, autonomous       quadrotor drones equipped with Neural-Fly learn how to respond to       strong winds so well that their performance significantly improved (as       measured by their ability to precisely follow a flight path). The error       rate following that flight path is around 2.5 times to 4 times smaller       compared to the current state of the art drones equipped with similar       adaptive control algorithms that identify and respond to aerodynamic       effects but without deep neural networks.              Neural-Fly, which was developed in collaboration with Caltech's       Yisong Yue, Professor of Computing and Mathematical Sciences,       and Anima Anandkumar, Bren Professor of Computing and Mathematical       Sciences, is based on earlier systems known as Neural-Lander and       Neural-Swarm. Neural-Lander also used a deep- learning method to track       the position and speed of the drone as it landed and modify its landing       trajectory and rotor speed to compensate for the rotors' backwash from       the ground and achieve the smoothest possible landing; Neural- Swarm       taught drones to fly autonomously in close proximity to each other.              Though landing might seem more complex than flying, Neural-Fly, unlike       the earlier systems, can learn in real time. As such, it can respond to       changes in wind on the fly, and it does not require tweaking after the       fact. Neural-Fly performed as well in flight tests conducted outside the       CAST facility as it did in the wind tunnel. Further, the team has shown       that flight data gathered by an individual drone can be transferred to       another drone, building a pool of knowledge for autonomous vehicles.              At the CAST Real Weather Wind Tunnel, test drones were tasked with       flying in a pre-described figure-eight pattern while they were blasted       with winds up to 12.1 meters per second -- roughly 27 miles per hour,       or a six on the Beaufort scale of wind speeds. This is classified as a       "strong breeze" in which it would be difficult to use an umbrella. It       ranks just below a "moderate gale," in which it would be difficult to       move and whole trees would be swaying. This wind speed is twice as fast       as the speeds encountered by the drone during neural network training,       which suggests Neural-Fly could extrapolate and generalize well to unseen       and harsher weather.              The drones were equipped with a standard, off-the-shelf flight control       computer that is commonly used by the drone research and hobbyist       community. Neural-Fly was implemented in an onboard Raspberry Pi 4       computer that is the size of a credit card and retails for around $20.                     ==========================================================================       Story Source: Materials provided by       California_Institute_of_Technology. Original written by Robert       Perkins. Note: Content may be edited for style and length.                     ==========================================================================       Journal Reference:        1. Michael O'Connell, Guanya Shi, Xichen Shi, Kamyar Azizzadenesheli,        Anima        Anandkumar, Yisong Yue, Soon-Jo Chung. Neural-Fly enables rapid        learning for agile flight in strong winds. Science Robotics, 2022;        7 (66) DOI: 10.1126/scirobotics.abm6597       ==========================================================================              Link to news story:       https://www.sciencedaily.com/releases/2022/05/220505085644.htm              --- up 9 weeks, 3 days, 10 hours, 50 minutes        * Origin: -=> Castle Rock BBS <=- Now Husky HPT Powered! (1:317/3)       SEEN-BY: 15/0 106/201 114/705 123/120 129/330 331 153/7715 218/700       SEEN-BY: 229/110 111 317 400 426 428 470 664 700 292/854 298/25 305/3       SEEN-BY: 317/3 320/219 396/45       PATH: 317/3 229/426           |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca