Controlling Drones with the Human Brain
A researcher from Arizona State University wants to command machines with the human brain. In fact, within ten years, Panagiotis Artemiadis, professor of mechanical and aerospace engineering at Arizona State University, envisions a swarm of brain-controlled drones playing a critical role in a range of applications, including agriculture and search-and-rescue operations.
The essential technology component of mind-controlled UAVs, or unmanned aerial vehicles, is the interface: a combination of hardware and algorithms that maps one’s brain activation to commands for a robotic system. Using electrodes placed on the scalp to record electrical activity, the non-invasive Electroencephalography (EEG) method measures voltage fluctuations resulting from the neurons’ ionic current. The hardware then measures brain activation as the subject thinks about an intended motion for the machine. The algorithms decode those activations to control commands for the robotic system.
At ASU’s Human-Oriented Robotics and Control (HORC) Lab, Artemiadis and his fellow researchers have evaluated the brain's capacity to direct non-human behaviors, including the flight of drones. In 2016, Artemiadis tested the ability of a single operator to control three UAVs, guiding each through a narrow loop. The subject used a hybrid brain-machine interface, combining EEG-recorded brain activity with input from a joystick. To control the multiple drones, the user had to think about shrinking or expanding the shape of the UAV grouping as a whole. By developing algorithms that extracted specific information from the brain, the experiment demonstrated the ability for humans to steer not just a single machine, but a swarm.
Collective behaviors are abundant in nature, but the idea of a flock or herd has only recently been adopted to robotic concepts. A robotic “swarm” system consists of a large group of interchangeable vehicles that use information, obtained via local sensing and communication, to execute autonomous decisions. Artemiadis and his team tested a “human-swarm interface” to, in effect, decode the brain and extract information related to desired collective behaviors, like the expansion of a swarm’s coverage or the implementation of a specific formation.
The algorithms developed at the HORC lab extract brain signals. The ASU team applies machine learning and pattern recognition algorithms to translate the brain’s electrical signals into control commands that are wirelessly transmitted to the drones.
The integration of very large teams of robots into comprehensive systems enables new tasks and missions, including search, exploration, and surveillance, according to the ASU researcher. Armed with infrared imaging equipment, a drone swarm, for example, could provide real-time tracking of a forest fire, allowing responders to adjust their plans on the fly. Teams of drones could also be called on to create topographic maps for soil analysis and irrigation planning.
Top Stories
INSIDERManned Systems
Turkey's KAAN Combat Aircraft Completes First Flight - Mobility Engineering...
INSIDERMaterials
FAA Expands Boeing 737 Investigation to Manufacturing and Production Lines -...
INSIDERImaging
New Video Card Enables Supersonic Vision System for NASA's X-59 Demonstrator -...
INSIDERManned Systems
Stratolaunch Approaches Hypersonic Speed in First Powered TA-1 Test Flight -...
INSIDERUnmanned Systems
Army Ends Future Attack and Reconnaissance Helicopter Development Program -...
ArticlesEnergy
Can Solid-State Batteries Commercialize by 2030? - Mobility Engineering...
Webcasts
AR/AI
From Data to Decision: How AI Enhances Warfighter Readiness
Energy
April Battery & Electrification Summit
Manufacturing & Prototyping
Tech Update: 3D Printing for Transportation in 2024
Test & Measurement
Building an Automotive EMC Test Plan
Manufacturing & Prototyping
The Moon and Beyond from a Thermal Perspective
Software
Mastering Software Complexity in Automotive: Is Release Possible...