dc.contributor.author |
Yacine Trad, Taha |
|
dc.contributor.author |
Choutri, Kheireddine |
|
dc.contributor.author |
Lagha, Mohand |
|
dc.contributor.author |
Khenfri, Fouad |
|
dc.date.accessioned |
2024-05-31T13:53:07Z |
|
dc.date.available |
2024-05-31T13:53:07Z |
|
dc.date.issued |
2024-05-31 |
|
dc.identifier.issn |
2210-142X |
|
dc.identifier.uri |
https://journal.uob.edu.bh:443/handle/123456789/5709 |
|
dc.description.abstract |
Unmanned Aerial Vehicles (UAVs), particularly quadrotors, have become highly versatile platforms for various applications
and missions. In this study, the employment of Multi-Agent Reinforcement Learning (MARL) in quadrotor control systems is investigated,
expanding its conventional usage beyond multi-UAV path planning and obstacle avoidance tasks. While traditional single-agent control
techniques face limitations in effectively managing the coupled dynamics associated with attitude control, especially when exposed to
complex scenarios and trajectories, this paper presents a novel method to enhance the adaptability and generalization capabilities of
Reinforcement Learning (RL) low-level control agents in quadrotors. We propose a framework consisting of collaborative MARL to
control the Roll, Pitch, and Yaw of the quadrotor, aiming to stabilize the system and efficiently track various predefined trajectories.
Along with the overall system architecture of the MARL-based attitude control system, we elucidate the training framework, collaborative
interactions among agents, neural network structures, and reward functions implemented. While experimental validation is pending,
theoretical analyses and simulations illustrate the envisioned benefits of employing MARL for quadrotor control in terms of stability,
responsiveness, and adaptability. Central to our approach is the employment of multiple actor-critic algorithms within the proposed
control architecture, and through a comparative study, we evaluate the performance of the advocated technique against a single-agent
RL controller and established linear and nonlinear methodologies, including Proportional-Integral-Derivative (PID) and Backstepping
control, highlighting the advantages of collaborative intelligence in enhancing quadrotor control in complex environments. |
en_US |
dc.language.iso |
en |
en_US |
dc.publisher |
University of Bahrain |
en_US |
dc.subject |
Quadrotors, Attitude Control, Multi-Agent Deep Reinforcement Learning, Collaborative Intelligence |
en_US |
dc.title |
Collaborative Multi-Agent Deep Reinforcement Learning Approach for Enhanced Attitude Control in Quadrotors |
en_US |
dc.identifier.doi |
http://dx.doi.org/10.12785/ijcds/1571033189 |
|
dc.volume |
17 |
en_US |
dc.issue |
1 |
en_US |
dc.pagestart |
1 |
en_US |
dc.pageend |
13 |
en_US |
dc.contributor.authorcountry |
Algeria |
en_US |
dc.contributor.authorcountry |
Algeria |
en_US |
dc.contributor.authorcountry |
Algeria |
en_US |
dc.contributor.authorcountry |
France |
en_US |
dc.contributor.authoraffiliation |
Aeronautical Sciences Laboratory, Aeronautical and Spatial Studies Institute, Blida 1 University |
en_US |
dc.contributor.authoraffiliation |
Aeronautical Sciences Laboratory, Aeronautical and Spatial Studies Institute, Blida 1 University |
en_US |
dc.contributor.authoraffiliation |
Aeronautical Sciences Laboratory, Aeronautical and Spatial Studies Institute, Blida 1 University |
en_US |
dc.contributor.authoraffiliation |
Energy and Embedded Systems for Transport, ESTACA’Lab |
en_US |
dc.source.title |
International Journal of Computing and Digital Systems |
en_US |
dc.abbreviatedsourcetitle |
IJCDS |
en_US |