dc.contributor.author | Puente-Castro, Alejandro | |
dc.contributor.author | Rivero, Daniel | |
dc.contributor.author | Pedrosa, Eurico | |
dc.contributor.author | Pereira, Artur | |
dc.contributor.author | Lau, Nuno | |
dc.contributor.author | Fernández-Blanco, Enrique | |
dc.date.accessioned | 2023-12-11T08:53:41Z | |
dc.date.available | 2023-12-11T08:53:41Z | |
dc.date.issued | 2023 | |
dc.identifier.citation | Puente-Castro, A., Rivero, D., Pedrosa, E., Pereira, A., Lau, N., & Fernandez-Blanco, E. (2023). Q-Learning based system for Path Planning with Unmanned Aerial Vehicles swarms in obstacle environments. Expert Systems With Applications, 235, 121240.https://doi.org/10.1016/j.eswa.2023.121240 | es_ES |
dc.identifier.uri | http://hdl.handle.net/2183/34437 | |
dc.description.abstract | [Abstract]: Path Planning methods for the autonomous control of Unmanned Aerial Vehicle (UAV) swarms are on the rise due to the numerous advantages they bring. There are increasingly more scenarios where autonomous control of multiple UAVs is required. Most of these scenarios involve a large number of obstacles, such as power lines or trees. Despite these challenges, there are also several advantages; if all UAVs can operate autonomously, personnel expenses can be reduced. Additionally, if their flight paths are optimized, energy consumption is reduced, leaving more battery time for other operations. In this paper, a Reinforcement Learning-based system is proposed to solve this problem in environments with obstacles by utilizing Q-Learning. This method allows a model, in this case, an Artificial Neural Network, to self-adjust by learning from its mistakes and successes. Regardless of the map’s size or the number of UAVs in the swarm, the goal of these paths is to ensure complete coverage of an area with fixed obstacles for tasks like field prospecting. Setting goals or having any prior information apart from the provided map is not required. During the experimentation phase, five maps of varying sizes were used, each with different obstacles and a varying number of UAVs. To evaluate the quality of the results, the number of actions taken by each UAV to complete the task in each experiment was considered. The results indicate that the system achieves solutions with fewer movements as the number of UAVs increases. An increasing number of UAVs on a map lead to solutions in fewer moves. The results have been compared, and a statistical significance analysis has been conducted on the proposed model’s outcomes, demonstrating its capabilities. Thus, it is shown that a two-layer Artificial Neural Network used to implement a Q-Learning algorithm is sufficient to operate on maps with obstacles. | es_ES |
dc.description.sponsorship | This project was supported by the FCT - Foundation for Science and Technology, Portugal, in the context of the project [grant number UIDB/00127/2020], and also POCI 2020, in the context of the Germirrad project [grant number POCI-01-0247-FEDER-072237]. Also, the General Directorate of Culture, Education, and University Management of Xunta de Galicia [grant number ED431D 2017/16]. This work was also funded by the grant for the consolidation and structuring of competitive research units [grant number ED431C 2022/46] from the General Directorate of Culture, Education and University Management of Xunta de Galicia, and the CYTED network, Spain [grant number PCI2018_093284] funded by the Spanish Ministry of Innovation and Science. This project was also supported by the General Directorate of Culture, Education and University Management of Xunta de Galicia “PRACTICUM DIRECT” [grant number IN845D-2020/03]. | es_ES |
dc.description.sponsorship | Portugal. Fundação para a Ciência e a Tecnologia; UIDB/00127/2020 | es_ES |
dc.description.sponsorship | Portugal. Programa Operacional Competitividade e Internacionalização; POCI-01-0247-FEDER-072237 | es_ES |
dc.description.sponsorship | Xunta de Galicia; ED431D 2017/16 | es_ES |
dc.description.sponsorship | Xunta de Galicia; ED431C 2022/46 | es_ES |
dc.description.sponsorship | Xunta de Galicia; IN845D-2020/03 | es_ES |
dc.language.iso | eng | es_ES |
dc.publisher | Elsevier | es_ES |
dc.relation | info:eu-repo/grantAgreement/AEI/Plan Estatal de Investigación Científica y Técnica y de Innovación 2017-2020/PCI2018_093284/ES/OBESIDAD Y DIABETES EN IBEROAMERICA: FACTORES DE RIESGO Y NUEVOS BIOMARCADORES PATOGENICOS Y PREDICTIVOS | es_ES |
dc.relation.uri | https://doi.org/10.1016/j.eswa.2023.121240 | es_ES |
dc.rights | Attribution-NonCommercial-NoDerivs 4.0 International (CC BY-NC-ND) | es_ES |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/3.0/es/ | * |
dc.subject | UAV | es_ES |
dc.subject | Artificial Neural Network | es_ES |
dc.subject | Reinforcement learning | es_ES |
dc.subject | Path Planning | es_ES |
dc.subject | Obstacle | es_ES |
dc.subject | Swarm | es_ES |
dc.title | Q-Learning based system for Path Planning with Unmanned Aerial Vehicles swarms in obstacle environments | es_ES |
dc.type | info:eu-repo/semantics/article | es_ES |
dc.rights.access | info:eu-repo/semantics/openAccess | es_ES |
UDC.journalTitle | Expert Systems with Applications | es_ES |
UDC.volume | 235 | es_ES |
UDC.issue | 121240 | es_ES |
dc.identifier.doi | 10.1016/j.eswa.2023.121240 | |
UDC.coleccion | Investigación | |
UDC.departamento | Ciencias da Computación e Tecnoloxías da Información | |
UDC.grupoInv | Redes de Neuronas Artificiais e Sistemas Adaptativos -Informática Médica e Diagnóstico Radiolóxico (RNASA - IMEDIR) | |