伤员转运后送
01-从角色2向角色3医疗设施航空医疗后送期间的战斗伤亡管理
03-Collective aeromedical evacuations of SARS-CoV-2-related ARDS patients in a military tactical plane- a retrospective descriptive study
04-乌克兰火车医疗后送的特点,2022
02-Decision Support System Proposal for Medical Evacuations in Military Operations
02-军事行动中医疗后送的决策支持系统建议
05-无人驾驶飞机系统的伤员疏散需要做什么
04-Characteristics of Medical Evacuation by Train in Ukraine, 2022.
05-Unmanned Aircraft Systems for Casualty Evacuation What Needs to be Done
07-一个德语语料库,用于搜索和救援领域的语音识别
08-雷达人类呼吸数据集的应用环境辅助生活和搜索和救援行动
08-Radar human breathing dataset for applications of ambient assisted living and search and rescue operations
06-基于信息融合的海上搜索救援目标定位
07-RESCUESPEECH- A GERMAN CORPUS FOR SPEECH RECOGNITION IN SEARCH AND RESCUE DOMAIN
12-欧盟和世卫组织联手进一步加强乌克兰的医疗后送行动
09-战场伏击场景下无人潜航器最优搜索路径规划
11-麦斯卡尔医疗后送-康涅狄格州陆军警卫医务人员在大规模伤亡训练中证明了他们的能力
06-Target localization using information fusion in WSNs-based Marine search and rescue
13- 年乌克兰火车医疗后送的特点
09-Optimal search path planning of UUV in battlefeld ambush scene
10-志愿医护人员从乌克兰前线疏散受伤士兵
14-海上搜救资源配置的多目标优化方法——在南海的应用
14-A Multi-Objective Optimization Method for Maritime Search and Rescue Resource Allocation An Application to the South China Sea
15-基于YOLOv5和分层人权优先的高效无人机搜索路径规划方法
17-乌克兰医疗保健专业人员在火药行动期间的经验对增加和加强培训伙伴关系的影响
17-Ukrainian Healthcare Professionals Experiences During Operation Gunpowder Implications for Increasing and Enhancing Training Partnerships
15-An Integrated YOLOv5 and Hierarchical Human Weight-First Path Planning Approach for Efficient UAV Searching Systems
16-基于旋转变压器的YOLOv5s海上遇险目标检测方法
16-YOLOv5s maritime distress target detection method based on swin transformer
19-人工智能的使用在伤员撤离、诊断和治疗阶段在乌克兰战争中
19-THE USE OF ARTIFICIAL INTELLIGENCE AT THE STAGES OF EVACUATION, DIAGNOSIS AND TREATMENT OF WOUNDED SOLDIERS IN THE WAR IN UKRAINE
18-军事行动中医疗后送的决策支持系统建议
20-乌克兰医疗保健专业人员在火药行动中的经验对增加和加强培训伙伴关系的影响
20-Ukrainian Healthcare Professionals Experiences During Operation Gunpowder Implications for Increasing and Enhancing Training Partnerships
21-大国冲突中医疗后送的人工智能
18-Decision Support System Proposal for Medical Evacuations in Military Operations
23-伤亡运输和 疏散
24-某军用伤员疏散系统仿真分析
23-CASUALTY TRANSPORT AND EVACUATION
24-Simulation Analysis of a Military Casualty Evacuation System
25-无人驾驶飞机系统的伤员疏散需要做什么
26-Aeromedical Evacuation, the Expeditionary Medicine Learning Curve, and the Peacetime Effect.
26-航空医疗后送,远征医学学习曲线,和平时期的影响
25-Unmanned Aircraft Systems for Casualty Evacuation What Needs to be Done
28-军用战术飞机上sars - cov -2相关ARDS患者的集体航空医疗后送——一项回顾性描述性研究
27-乌克兰火车医疗后送的特点,2022
27-Characteristics of Medical Evacuation by Train in Ukraine, 2022.
28-Collective aeromedical evacuations of SARS-CoV-2-related ARDS patients in a military tactical plane- a retrospective descriptive study
03-军用战术飞机上sars - cov -2相关ARDS患者的集体航空医疗后送——一项回顾性描述性研究
30-评估局部现成疗法以减少撤离战场受伤战士的需要
31-紧急情况下重伤人员的医疗后送——俄罗斯EMERCOM的经验和发展方向
31-Medical Evacuation of Seriously Injured in Emergency Situations- Experience of EMERCOM of Russia and Directions of Development
30-Evaluation of Topical Off-the-Shelf Therapies to Reduce the Need to Evacuate Battlefield-Injured Warfighters
29-军事行动中医疗后送的决策支持系统建议
29-Decision Support System Proposal for Medical Evacuations in Military Operations
32-决策支持在搜救中的应用——系统文献综述
32-The Syrian civil war- Timeline and statistics
35-印尼国民军准备派飞机接运 1
33-eAppendix 1. Information leaflet basic medical evacuation train MSF – Version April 2022
36-战场上的医疗兵
34-Characteristics of Medical Evacuation by Train in Ukraine
22-空军加速变革以挽救生命:20年来航空医疗后送任务如何取得进展
34-2022年乌克兰火车医疗疏散的特点
33-信息传单基本医疗后送车
40-航空医疗后送
43-美军的黄金一小时能持续多久
42-陆军联手直升机、船只和人工智能进行伤员后送
47-受伤的士兵撤离
46-伤员后送的历史从马车到直升机
37-从死亡到生命之路
41-后送医院
52-印度军队伤员航空医疗后送经验
53-“地狱之旅”:受伤的乌克兰士兵撤离
45-伤病士兵的撤离链
54-热情的和资源匮乏的士兵只能靠自己
57-2022 年乌克兰火车医疗后送
51-医务人员在激烈的战斗中撤离受伤的乌克兰士兵
59-乌克兰展示医疗后送列车
61-俄罗斯士兵在乌克兰部署自制UGV进行医疗后送
60-“流动重症监护室”:与乌克兰顿巴斯战斗医务人员共24小时
50-医疗后送——保证伤员生命安全
阿拉斯加空军国民警卫队医疗后送受伤陆军伞兵
航空撤离,印度经验 抽象的
通过随机森林模拟规划方法解决军事医疗后送问题
2022 年乌克兰火车医疗后送的特点
战术战地救护教员指南 3E 伤员后送准备和要点 INSTRUCTOR GUIDE FOR TACTICAL FIELD CARE 3E PREAPRING FOR CASUALTY EVACUTION AND KEY POINTS
军事医疗疏散
北极和极端寒冷环境中的伤亡疏散:战术战斗伤亡护理中创伤性低温管理的范式转变
-外地伤员后送现场伤亡疏散
伤员后送图片
从角色2到角色3医疗设施期间战斗人员伤亡管理
关于军事行动中医疗疏散的决策支持系统建议书
在军事战术平面上对sars-cov-2相关 ARDS患者进行的集体空中医疗后送: 回顾性描述性研究
2022年乌克兰火车医疗疏散的特点
透过战争形势演变看外军营救后送阶梯 及医疗救护保障措施
东部伤兵营 英文 _Wounded_Warrior_Battalion_East
组织紧急医疗咨询和医疗后送 2015 俄文
-
+
首页
15-An Integrated YOLOv5 and Hierarchical Human Weight-First Path Planning Approach for Efficient UAV Searching Systems
<p><a href="https://www.mdpi.com/journal/machines"><img src="/media/202408//1724838585.8174589.jpeg" /><img src="/media/202408//1724838585.829876.png" /><img src="/media/202408//1724838585.8372521.png" /><img src="/media/202408//1724838585.841884.png" /><strong><em> machines</em></strong></a></p><p><img src="/media/202408//1724838585.864337.png" /><img src="/media/202408//1724838585.94442.png" /><a href="https://www.mdpi.com">mie,</a></p><p>Article</p><p><strong>An Integrated YOLOv5 and Hierarchical Human-Weight-First Path Planning Approach for Efficient UAV Searching Systems</strong></p><p><strong>Ing-Chau Chang 1,*</strong><a href="https://orcid.org/0000-0001-9088-9963"><img src="/media/202408//1724838585.9763799.png" /></a><strong>, Chin-En Yen 2</strong><a href="https://orcid.org/0000-0002-9760-936X"><img src="/media/202408//1724838586.043723.png" /></a><strong>, Hao-Fu Chang 1, Yi-Wei Chen1, Ming-Tsung Hsu 1, Wen-Fu Wang 1, Da-Yi Yang 1 and Yu-Hsuan Hsieh 1</strong></p><p><img src="/media/202408//1724838586.070845.png" /><a href="https://www.mdpi.com/article/10.3390/machines12010065?type=check_update&version=1">巴</a></p><p><a href="https://www.mdpi.com/article/10.3390/machines12010065?type=check_update&version=1">check for</a></p><p><a href="https://www.mdpi.com/article/10.3390/machines12010065?type=check_update&version=1">updates</a></p><p><strong>Citation: </strong>Chang,I.-C.; Yen, C.-E.;</p><p>Chang, H.-F.; Chen, Y.-W.; Hsu, M.-T.; Wang, W.-F.; Yang, D.-Y.; Hsieh, Y.-H. An Integrated YOLOv5 and</p><p>Hierarchical Human-Weight-First</p><p>Path Planning Approach for Efficient UAV Searching Systems. Machines</p><p><strong>2024</strong>, 12, 65. <a href="https://doi.org/10.3390/machines12010065">https://doi.org/</a> <a href="https://doi.org/10.3390/machines12010065">10.3390/machines12010065</a></p><p>Academic Editors: Maria</p><p>Eusebia Guerrero-Sanchez and Omar Hernández-González</p><p>Received: 3 December 2023 Revised: 6 January 2024</p><p>Accepted: 8 January 2024 Published: 16 January 2024</p><p><a href="https://creativecommons.org/"><img src="/media/202408//1724838586.2176409.png" /></a></p><p><strong>Copyright: </strong>© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license <a href="https://creativecommons.org/licenses/by/4.0/">(https://</a> <a href="https://creativecommons.org/licenses/by/4.0/">creativecommons.org/licenses/by/</a> 4.0/).</p><p>1 Department of Computer Science and Information Engineering, National Changhua University of Education,</p><p>Changhua 50007, Taiwan; fchang@wisdome.ai (H.-F.C.); yiwei.chen@g.ncu.edu.tw (Y.-W.C.); xmc510063@gapp.nthu.edu.tw (M.-T.H.); wang1020351@gapp.nthu.edu.tw (W.-F.W.);</p><p>m1254004@gm.ncue.edu.tw (D.-Y.Y.); s0954010@mail.ncue.edu.tw (Y.-H.H.)</p><p>2 Department of Early Childhood Development and Education, Chaoyang University of Technology, Taichung 41349, Taiwan; ceyen@cyut.edu.tw</p><p><strong>* </strong>Correspondence: icchang@cc.ncue.edu.tw; Tel.: +886-4-723-2105</p><p><strong>Abstract: </strong>Because the average number of missing people in our country is more than 20,000 per year, determining how to efficiently locate missing people is important. The traditional method of finding missing people involves deploying fixed cameras in some hotspots to capture images and using humans to identify targets from these images. However, in this approach, high costs are incurred in deploying sufficient cameras in order to avoid blind spots, and a great deal of time and human effort is wasted in identifying possible targets. Further, most AI-based search systems focus on how to improve the human body recognition model, without considering how to speedup the search in order to shorten the search time and improve search efficiency, which is the aim of this study. Hence, by exploiting the high-mobility characteristics of unmanned aerial vehicles (UAVs), this study proposes an integrated YOLOv5 and hierarchical human-weight-first (HWF) path planning framework to serve as an efficient UAV searching system, which works by dividing the whole searching process into two levels. At level one, a searching UAV is dispatched to a higher altitude to capture images, covering the whole search area. Then, the well-known artificial intelligence model YOLOv5 is used to identify all persons in the captured images and compute corresponding weighted scores for each block in the search area, according to the values of the identified human bodies, clothing types, and clothing colors. At level two, the UAV lowers its altitude to sequentially capture images for each block, in descending order according to its weighted score at level one, and it uses the YOLOv5 recognition model repeatedly until the search target is found. Two improved search algorithms, HWFR-S and HWFR-D, which incorporate the concept of the convenient visit threshold and weight difference, respectively, are further proposed to resolve the issue of the lengthy and redundant flight paths of HWF. The simulation results suggest that the HWF, HWFR-S, and HWFR-D search algorithms proposed in this study not only effectively reduce the length of a UAV’s search path and the number of search blocks but also decrease the search time required for a UAV to locate the search target, with a much higher search accuracy than the two traditional search algorithms. Moreover, this integrated YOLOv5 and HWF framework is implemented and tested in a real scenario to demonstrate its capability in enhancing the efficiency of a search and rescue operation.</p><p><strong>Keywords: </strong>unmanned aerial vehicle; hierarchical human-weight-first path planning; artificial intelligence image recognition; YOLOv5; searching corners without cameras</p><p><strong>1. Introduction</strong></p><p>As the technology of the unmanned aerial vehicle (UAV) has seen significant progress in recent years, a number of applications have been proposed for it [<a href="#bookmark1">1,</a><a href="#bookmark2">2</a>] due to its unique characteristics, such as higher mobility and more flexible integration with different equip- ment, such as sensors and cameras, etc. <a href="#bookmark3">[3,</a><a href="#bookmark4">4]</a>. The researchers of [<a href="#bookmark5">5</a>] explore algorithms for</p><p>Machines <strong>2024</strong>, 12, 65. <a href="https://doi.org/10.3390/machines12010065">https://doi.org/10.3390/machines12010065</a> <a href="https://www.mdpi.com/journal/machines">https://www.mdpi.com/journal/machines</a></p><p>the formation movement of UAV swarms, with the objective of facilitating simultaneous adjustments to the formation shape while the UAV swarm is in motion. Signal transmission is another highly significant topic in UAV control. The research in [<a href="#bookmark6">6</a>] proposes automatic modulation classification utilizing deep learning in this context. The study in [<a href="#bookmark7">7</a>] addresses improvements to existing GNSS systems, such as GPS positioning, tackling issues related to inaccuracies. The researchers propose a time-differenced carrier phase (TDCP) derivation- controlled GNSS/IMU integration scheme to successfully acquire vehicle information such as the relative position and heading. Real-world tests demonstrate that this method exhibits higher accuracy compared to traditional algorithms. In recent years, the increas- ing integration of UAVs with various interdisciplinary domains has also been observed. Koopman operators are mathematical tools used to describe the evolution of nonlinear dynamic systems. The work of [<a href="#bookmark8">8</a>] proposes robust tube-based model predictive control with Koopman operators, while [<a href="#bookmark9">9</a>] integrates Koopman operators with the control of UAVs. Furthermore, there exist various UAV path planning problems and related studies, such as the capacitated arc routing problem (CARP). The objective of CARP is to find the shortest path in a mixed graph with undirected edges and directed arcs, minimizing the distance of the path while considering capacity constraints for objects moving on the graph. In [<a href="#bookmark10">10</a>], the study introduces a memetic algorithm based on Two_Arch2 (MATA),which simultaneously considers multiple optimization objectives for the path planning problem, including the total cost, makespan, carbon emissions, and load utilization rate.</p><p>Recently, UAVs have been used for search and rescue (SAR) missions to find missing persons at the scene of a natural disaster or when an emergency event occurs [<a href="#bookmark11">11</a>–<a href="#bookmark12">13]</a>. The issue of missing persons is a challenging societal problem, particularly when involving minors. Children, due to their smaller stature, are susceptible to disappearance within large crowds, especially in crowded places such as amusement parks, making it difficult to notice their absence. Unfortunately, they generally exhibit a lower level of vigilance towards unfamiliar individuals, rendering them vulnerable to abduction. As the duration of a missing person’s search is prolonged, the probability of encountering a perilous situation escalates, imposing significant psychological distress upon parents.</p><p>However, there is a limited amount of research aimed at identifying specific indi- viduals, such as missing persons, and researchers have primarily relied on fixed cameras installed in specific areas. This limitation prevents the continuous tracking of targets, leading to difficulties in inferring their actual positions due to the limited perspective and potential blind spots. Furthermore, most of the existing works on search and rescue adopt unmanned aerial vehicles (UAVs) [<a href="#bookmark13">14</a>–<a href="#bookmark14">16</a>] and employ indiscriminate search algorithms, without prioritizing the areas where the search target maybe located, resulting in inefficient search operations and excessive UAV power consumption. Hence, intelligent approaches such as AI-enabled object detection in UAVs [<a href="#bookmark15">17</a>] are proposed to overcome this problem. One of the major research directions regarding SAR involves coordinating multiple UAVs to cover multiple regions, which maybe scattered or not [<a href="#bookmark16">18</a>–<a href="#bookmark17">21]</a>. These approaches focus on how to balance the time consumption of UAVs or design effective coverage flights to cover the search area completely, with no collisions of multiple UAVs. On the other hand, this study aims to improve the search time and costs related to an assigned search area for a single UAV. It addresses the inefficiency problem of SAR mentioned above by proposing an integrated YOLOv5 and hierarchical HWF path planning approach. This integrated approach executes the following steps. First, the user specifies the geographical coordinates of the search area and important features of the search target. Then, the searching UAV is dispatched to a higher altitude to capture images covering the whole search area using an onboard camera. Next, the search area is partitioned into multiple blocks. The Jetson Nano mounted on the searching UAV captures and transmits images to a cloud server through wireless communication. On the cloud server, the well-known artificial intelligence model YOLOv5 is used for human body recognition and clothing recognition, and the KNN algorithm is used to identify clothing colors. Corresponding weighted scores are computed with the identified human bodies, clothing types, and clothing colors within each</p><p>block. After this, the cloud server issues a command to the UAV to adjust its coordinates and lower its altitude to capture clearer images that cover a single block. The proposed human-weight-first (HWF) path planning algorithm is utilized to guide the UAV to visit blocks sequentially, according to their weighted scores, in descending order. YOLOv5 and the KNN algorithm are iteratively employed on the cloud server side to recognize human bodies, clothing types, and clothing colors for images captured at lower altitudes. Upon confirming the presence of the search target, the human body image and its location are reported to the user, which concludes the UAV’s search mission. This integrated YOLOv5 and hierarchical HWF path planning approach can prioritize the block where the search target is most likely to be located, to shorten the search time and costs, which significantly improves the search efficiency and avoids the searching of corners without cameras.</p><p>Consequently, this study achieves the following contributions.</p><p>1. It utilizes the existing YOLOv5 model to automatically recognize the search target and uses KNN color recognition to recognize clothing/pant colors in realtime; it thus avoids wasting time and human effort in the manual identification of possible targets.</p><p>2. According to the recognition results of the YOLOv5 model and KNN color recognition, the study proposes a weighting subroutine to calculate the human weights of each block and the hierarchical human-weight-first (HWF) path planning algorithm to dispatch the UAV to capture images repeatedly of the search area and each block at different altitudes.</p><p>3. It proposes a complete flowchart of the integrated YOLOv5 and HWF framework to reduce the search time and avoid searching corners without cameras.</p><p><strong>2. Related Work</strong></p><p>2.1. Traditional Unmanned Aerial Vehicle Path Planning Methods for Search and Rescue Operations</p><p>Several search and rescue methods have been proposed recently [<a href="#bookmark13">14</a>–<a href="#bookmark14">16]</a>. In [<a href="#bookmark13">14</a>], the sweep line search method conducts a thorough search from left to right, as illustrated in Figure <a href="#bookmark18">1.</a> Meanwhile, ref. <a href="#bookmark19">[15</a>] introduces the spiral search, which navigates the designated search area in a spiral pattern, as depicted in Figure <a href="#bookmark20">2.</a> Both methods are uncomplicated and exhibit algorithms with linear time complexity in relation to the search area. Differing from these two methods, refs. <a href="#bookmark14">[16,</a><a href="#bookmark21">22</a>] introduce block-based methods. These approaches offer the advantage of categorizing the whole search area into blocks with and without search targets. Figure <a href="#bookmark22">3</a>demonstrates the relationship between the UAV’s perspective and the altitude concerning the search blocks when the whole search area is partitioned [<a href="#bookmark21">22]</a>. Through the traveling salesman problem (TSP) [<a href="#bookmark23">23</a>] approach, the shortest path that does not require the visiting of all blocks is computed if all blocks with search targets have been recognized in advance. However, the four methods mentioned above do not prioritize the block searching sequence in proximity to the search target, which results in inadequate search efficiency. Therefore, taking inspiration from block-based approaches, this study assigns priority to all blocks based on the likelihood of the blocks containing potential targets, which are automatically recognized in real time using the YOLOv5 model. In contrast to [<a href="#bookmark14">16</a>], which primarily focuses on finding the shortest path, this study emphasizes improving the search efficiency to yield the shortest search time by searching in the block <a id="bookmark18"></a>with the highest priority first.</p><p><img src="/media/202408//1724838586.5362718.png" /></p><p><strong>Figure 1. </strong>Sweep line search.</p><p><img src="/media/202408//1724838586.598772.png" /></p><p><a id="bookmark20"></a><a id="bookmark22"></a><strong>Figure 2. </strong>Spiral search.</p><p><img src="/media/202408//1724838586.678733.png" /></p><p><strong>Figure 3. </strong>The relationship between the altitude of the UAV and the partitioned search area.</p><p>2.2. Search Target Recognition Techniques</p><p>2.2.1. Color Space Exchange</p><p>The RGB color space is the most widely used color space, where RGB denotes red, green, and blue. It is similar to the well-known concept of the primary colors of light, where mixing these colors yields various levels of brightness and chromaticity. However, the RGB color space has a strong dependence on the lighting conditions, meaning that the color of an object can change with variations in brightness. In addition, the three elements in the RGB color space are highly correlated, indicating that a change in one element will result in a corresponding change in the perceived color. Therefore, using the RGB color space for the color extraction of objects is not ideal [<a href="#bookmark24">24]</a>. In contrast, the HSV color space [<a href="#bookmark25">25</a>] is more intuitive and easily understood compared to the RGB color space. It separates the brightness value (V) from the color chrominance, which can be further divided into hue (H) and saturation (S). Because these elements in HSV have a relatively weak correlation with each other, it is highly suitable for use in feature color extraction. In comparison to RGB, one of the advantages of the HSV color space is its weak inter-element correlation, making it easy to control. In applications involving color recognition, we can convert the detected images from the RGB color space to the HSV color space with Equation (1).</p><p><img src="/media/202408//1724838586.804846.png" /></p><p><img src="/media/202408//1724838587.132419.png" /></p><p>2.2.2. Extracting Feature Colors of Image</p><p>The feature color extraction process in [<a href="#bookmark25">25</a>] involves first segmenting the elements of an image’s HSV color space, followed by the conversion of each element (H, S, V) into a histogram of oriented gradient (HOG). Since the HOG divides each element into several element intervals, the segmentation proportions for each element can be determined. Then, selecting the interval with the highest proportion for each element, we can obtain their respective numerical values (H, S, V). These values represent the HSV feature colors for the image.</p><p>2.2.3. Transformation of Color Space</p><p>After experimenting, it has been observed that certain issues exist when directly calculating color distances in the HSV color space. Specifically, when the saturation (S) is low, it often leads to the k-nearest neighbors (KNN) [<a href="#bookmark26">26</a>] decision result being mistakenly classified as gray, regardless of how the hue (H) changes. To address this, the extracted feature colors in HSV are transformed into the RGB color space using Equation (2) [<a href="#bookmark25">25]</a>. This transformation involves mapping the hue (h) range to hi, and calculating variables p, q, t based on the hue (hi) range to determine which combination of RGB attributes (p, q, t, v) applies. The calculated RGB values (r0, g0, b0) are then subjected to Euclidean distance computation [<a href="#bookmark27">27</a>] against pre-established RGB color table values (r1, g1, b1) to determine the color distance (d), as illustrated in Equation (3). Subsequently, the KNN algorithm is employed to identify the color of the clothing based on this computed distance.</p><table><tr><td><p>hi = <img src="/media/202408//1724838587.186143.png" /></p><p>f = <img src="/media/202408//1724838587.1907232.png" /></p><p>p = v × (1 − s)</p><p>q = v × (1 − f × s)</p><p>t = v × (1 − (1 − f) × s)</p><p><img src="/media/202408//1724838587.1960058.png" /></p></td><td><p>(2)</p></td></tr><tr><td></td><td></td></tr><tr><td><p>d = √ (r1 − r0)2 + (g1 − g0)2 + (b1 − b0 )2,</p></td><td><p>(3)</p></td></tr></table><p>2.2.4. K-Nearest Neighbors (KNN) Color Classification</p><p>K-nearest neighbors (KNN) [<a href="#bookmark26">26</a>] is a fundamental classification and regression al- gorithm. After obtaining the HSV feature colors of an image and calculating the color distances using Equation (3), these distances are compared to a pre-established RGB color table. After sorting the color distances for each color, K colors with the closest distances are then selected. Followed by a voting process among neighboring colors, the color with the most votes is determined as the final color resultselected by the KNN algorithm.</p><p>2.2.5. UAV Systems for Human Detection</p><p>The work in [<a href="#bookmark28">28</a>] proposes an approach utilizing an automated human detection system on UAVs to identify human bodies, discussing the hardware configuration of UAVs and real-time human recognition capabilities. Ref. <a href="#bookmark29">[29</a>] presents a comprehensive human activity recognition algorithm, where the UAV first identifies whether the object is a person and subsequently recognizes various human activities, such as throwing, walking, and digging. Additionally, the study introduces various image stabilization techniques. The research of [<a href="#bookmark19">15</a>] focuses on achieving human body recognition using a CNN. Due to the difficulty in acquiring datasets, data augmentation is employed to enhance the training outcomes. The study compares the training outcomes using various architectures and outlines the algorithm’s path planning as a spiral search. The focus of the study in [<a href="#bookmark30">30]</a> lies in the application of UAVs for commercial transportation, aiming to achieve successful human body recognition using UAVs. The research encompasses the design of five distinct scenarios, revealing that the distance variation between the UAV and the human body has a more significant impact on the recognition success compared to the quality of the camera. In the context of search and rescue operations for swimmers, ref. <a href="#bookmark31">[31</a>] proposes a methodology that integrates global navigation satellite system (GNSS) techniques with computer vision algorithms to locate individuals in distress. Refs. <a href="#bookmark32">[32,</a><a href="#bookmark33">33</a>] primarily focus</p><p>on the training of human detection models. Ref. <a href="#bookmark32">[32</a>] introduces a modified YOLOv8 architecture by incorporating the SC3T module into the final layer and training the model using images captured from a UAV perspective. The emphasis of the study lies in the recognition performance. The experimental results are evaluated using confusion matrices and the mean average precision. The findings reveal that, across the precision rate, recall rate, and mAP, the modified YOLOv8 outperforms both the original YOLOv5 and YOLOv8 models. Ref. <a href="#bookmark33">[33</a>] primarily utilizes YOLOv5 for human detection and further employs a Haar cascade classifier to identify specific body parts (head, upper body, lower body). The final results indicate that YOLOv5 achieves 98% average precision (AP), while the Haar cascade classifier attains approximately 78% AP. Table <a href="#bookmark34">1</a>presents a comparison of related studies on human detection using UAVs. It can be found that most of the related methods focus on how to improve the human body recognition model, without considering how to speed up the search in order to shorten the search time and search efficiency, which is the aim of this study. Hence, the integrated YOLOv5 and HWF framework is proposed here to obtain an efficient UAV searching system by combining the hierarchical human-weight-first (HWF) path planning algorithm with the results of human body recognition from the</p><p>existing YOLOv5 model and the clothing/pant colors from KNN color recognition. <a id="bookmark34"></a><strong>Table 1. </strong>Comparison of related studies of UAV human detection.</p><table><tr><td></td><td><p><strong>Human Body</strong></p><p><strong>Recognition Model</strong></p></td><td><p><strong>Dataset Used</strong></p></td><td><p><strong>Recognition of Human Clothing Types and Colors</strong></p></td><td><p><strong>Segmentation</strong></p><p><strong>of the Search</strong></p><p><strong>Area</strong></p></td><td><p><strong>Dynamic Route Planning</strong></p><p><strong>for Search</strong></p></td><td><p><strong>Integration of</strong></p><p><strong>Human Body and</strong></p><p><strong>Clothing/Pant Color</strong></p><p><strong>Recognition with</strong></p><p><strong>Dynamic</strong></p><p><strong>Route Planning</strong></p></td></tr><tr><td><p><a href="#bookmark28">[28]</a></p></td><td><p>Motion detection outputs a score of human confidence</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td></tr><tr><td><p><a href="#bookmark29">[29]</a></p></td><td><p>CNN</p></td><td><p>UCF-ARG dataset</p></td><td><p>No, proposes human</p><p>activity classifica-</p><p>tion algorithm</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td></tr><tr><td><p><a href="#bookmark19">[15]</a></p></td><td><p>CNN</p></td><td><p>Self-developed captured dataset</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No, spiral search</p></td><td><p>No</p></td></tr><tr><td><p><a href="#bookmark30">[30]</a></p></td><td><p>DNN with MobileNet V2 SSDLite</p></td><td><p>COCO dataset</p></td><td><p>No</p></td><td><p>No</p></td><td><p>Yes, estimates the</p><p>person and moves</p><p>in his direction</p><p>with GPS</p></td><td></td></tr><tr><td><p><a href="#bookmark31">[31]</a></p></td><td><p>CNN with</p><p>Tiny YOLOv3</p></td><td><p>COCO</p><p>dataset + self-</p><p>developed</p><p>swimmers dataset</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td></tr><tr><td><p><a href="#bookmark32">[32]</a></p></td><td><p>CNN with</p><p>modified YOLOv8</p></td><td><p>Self-developed</p><p>UAV view</p><p>real-world dataset</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td></tr><tr><td><p><a href="#bookmark33">[33]</a></p></td><td><p>CNN with YOLOv5</p><p>and Haar</p><p>Cascade classifier</p></td><td><p>VisDrone dataset + COC0128 dataset</p></td><td><p>No, proposes a</p><p>human body region classification algorithm</p></td><td><p>No</p></td><td><p>No</p></td><td><p>No</p></td></tr><tr><td><p>HWF</p></td><td><p>CNN with YOLOv5</p></td><td><p>VisDrone dataset</p><p>+ self-developed</p><p>drone-</p><p>clothing dataset</p></td><td><p>Yes, uses KNN color recognition</p></td><td><p>Yes</p></td><td><p>Yes, proposes the</p><p>hierarchical</p><p>human-weight-first</p><p>(HWF) path</p><p>planning algorithm</p></td><td><p>Yes,</p><p>Proposes the</p><p>integrated YOLOv5</p><p>and HWF framework</p></td></tr></table><p><strong>3. System Architecture and Algorithms</strong></p><p>3.1. System Architecture</p><p>As depicted in Figure <a href="#bookmark35">4</a>, the system architecture is divided into three main components: the UAV equipped with the Jetson Nano [<a href="#bookmark36">34</a>], the server side, and the client side, where the Jetson Nano handles the UAV’s flight commands and mobile network communication</p><p>tasks. The search begins with the user inputting the location and the block size of the search <a id="bookmark35"></a>area, as well as four target searching criteria, into the server.</p><p><img src="/media/202408//1724838587.246339.png" /></p><p><strong>Figure 4. </strong>System architecture and flowchart of this study.</p><p>The four target searching conditions are shown below:</p><p>• Clothing type;</p><p>• Pant type;</p><p>• Clothing color;</p><p>• Pant color.</p><p>After the inputs, the server establishes a communication link with the Jetson Nano on the UAV via a mobile network connection, and the UAV is then dispatched from the starting point to the search area to initialize the search operation. Using a USB camera mounted on the UAV, aerial images are captured for the search area and blocks. All captured images and their shooting locations are sent back to the server, which benefits from its enhanced computational ability to undertake more intricate operations, including human body recognition by YOLOv5, clothing recognition by YOLOv5, and color identification by the KNN algorithm. This setup harnesses the advantages of cloud computing [<a href="#bookmark37">35]</a>. During the UAV’s searching process according to the hierarchical human-weight-first (HWF) path planning algorithm, if the server identifies a search target that meets the search criteria, it notifies the user and the Jetson Nano on the UAV. This instruction prompts the UAV to return to the starting point, concluding its search mission.</p><p>3.2. Search Algorithm</p><p>3.2.1. Hierarchical Flight Altitudes for the UAV</p><p>In this study, the UAV’s flight altitude is divided into two types: the initial altitude h0 to capture the image of the whole search area and altitude h1 to capture an image of a single block to achieve the optimal object recognition. The server conducts path planning at initial altitude h0 and then directs the UAV to fly at the optimal object recognition altitude h1 to traverse the blocks. According to Equation (4), the server calculates altitude h0 by utilizing the side length of the search area, i.e., √area, and tan<em>θ</em>, where <em>θ </em>is half of the field of view (FOV) of the UAV’s camera. This calculation determines the UAV heighth0 at which the image captured by the UAV’s camera with this field of view can cover the entire search area.</p><p><img src="/media/202408//1724838587.2537.png" /> (4)</p><p>Subsequently, the optimal object recognition altitude h1 is determined by calculation based on the predefined side length n of the block. At heighth1, the UAV has a clearer</p><p>perspective to achieve better recognition results, enhancing the accuracy in identifying the search target. When the block size (n × n) remains constant, the number of total blocks increases as the search area becomes larger. Conversely, as the size of the search area decreases, fewer blocks are segmented, as illustrated in Figure <a href="#bookmark38">5.</a> After calculating the individual weight values for all blocks at initial altitude h0, the human-weight-first (HWF) algorithm is employed to plan the flight path at the block level. Once the flight pathis planned,the UAV descends from initial altitude h0 to altitude h1 to start the block search. During the search path of the block level, if the server identifies an object from the image captured by the UAV at altitude h1 with a weight score exceeding the predefined threshold for the search target,it is considered that the target has been found. In this case, the server <a id="bookmark38"></a>sends the target’s location and the corresponding captured image back to the user and this concludes the UAV’s search mission.</p><p><img src="/media/202408//1724838587.2594821.png" /></p><p><strong>Figure 5. </strong>The number of blocks versus the area size.</p><p>3.2.2. Block Weight in the Search Area</p><p>The weight within each block is determined by the accuracy value of the recognized person’s body (APha), the accuracy values of the clothing and pant types (APca, APpa ), the recognized clothing and pant types (Crct, Crpt), the search clothing and pant types (Csct, Cspt), the recognized clothing and pant colors (Crcc, Crpc ), and the search clothing and pant colors (Cscc, Cspc ) in a fuzzy manner. This calculation results in the weight value,i.e., human_weight, for the recognized person. The block weight,i.e., block_weight, is defined as the highest weight value among all recognized persons in a block. By sorting the block weight values in the whole search area, it is possible to determine which block is most likely to contain the search target.</p><p>3.2.3. Hierarchical Human-Weight-First (HWF) Path Planning Algorithm</p><p>Traditional research such as [<a href="#bookmark37">35</a>] proposes the exhaustive approach to generate all possible paths,select the shortest path in terms of the pathlength, and finally traverse the blocks in descending order of the block weight. However, the exhaustive algorithm exhibits exponential time complexity as the size of the search area increases. Therefore, in this study, the human-weight-first (HWF) algorithm is designed as the path planning algorithm for the UAV. The HWF algorithm makes the optimal choice at each step in the current state, as in [<a href="#bookmark39">36</a>], aiming to achieve an overall result that is as optimal as possible.</p><p>The flow of the HWF algorithm is listed below and illustrated in Figure <a href="#bookmark40">6.</a></p><p>(a) The UAV flies from the center point of the search area to an altitude of h0 to begin</p><p>recognition, which is shown in Figure <a href="#bookmark40">6</a>a. Block weights are calculated using the image captured by the UAV at altitude h0 .</p><p>(b) The HWF algorithm selects the block with the highest block weight as the starting point for the block search and guides the UAV to descend to the center of the block at</p><p>altitude h1, which is shown in Figure <a href="#bookmark40">6</a>b. The UAV then captures images of the block and sends them to the server for further recognition.</p><p>(c) If no search target is found in this block, HWF instructs the UAV to traverse to the block with the next highest block weight until the search target is found,i.e., the block weight exceeds the search target threshold, or all blocks with nonzero block weights <a id="bookmark40"></a>have been visited, as shown in Figure <a href="#bookmark40">6c</a>.</p><p><img src="/media/202408//1724838587.2781608.png" /></p><p>(<strong>a</strong>)</p><p><img src="/media/202408//1724838587.290272.jpeg" /></p><p>(<strong>b</strong>)</p><p><img src="/media/202408//1724838587.320148.jpeg" /></p><p>(<strong>c</strong>)</p><p><strong>Figure 6. </strong>The complete planned path by HWF. (<strong>a</strong>) The UAV flies from the center point of the search area to an altitude of h0 to begin recognition. (<strong>b</strong>) The HWF algorithm selects the block with the highest block weight as the starting point for the block search and guides the UAV to descend to the center of the block at altitude h1 . (<strong>c</strong>) The red lines show the complete block traversal order.</p><p>As depicted in Figure <a href="#bookmark40">6</a>, the value in a block represents its block weight, which is used to prioritize the block search order, increasing the speed when identifying the search target. Hence, the complete block traversal order in Figure <a href="#bookmark40">6</a>c is as follows: block [0.9] → block [0.85] → block [0.74] → block [0.49] → block [0.58] → block [0.3]. It should be noted that the proposed integrated YOLOv5 and HWF approach results in a high probability of</p><p>recognizing the search target in the first few blocks, which significantly reduces the UAV search time, traversal distance, and power consumption.</p><p><a id="bookmark41"></a>3.2.4. Convenient Visit Algorithms Based on HWF</p><p>In Figure <a href="#bookmark42">7</a>, the black values represent the block weights calculated at altitude h0, while the blue values represent the weight scores calculated at altitude h1 . Assume that the target threshold is set as 80. Since the weights at both h0 and h1 altitudes do not exceed the target threshold, the HWF path planning algorithm performs a complete search over all blocks with nonzero weights,i.e., from the block with weight 76.0 to that with weight 5.1. This means that the HWF algorithm may lead to a UAV flight path that passes through some intermediate blocks several times, which increases the search pathlength, search delay, and UAV power consumption accordingly. This situation intensifies significantly when the search area expands. To address this issue, we design two variants of the HWF algorithm to search these intermediate blocks when it is convenient, if certain conditions <a id="bookmark42"></a>are met,thus addressing the aforementioned problems.</p><p><img src="/media/202408//1724838587.3596082.jpeg" /></p><p><strong>Figure 7. </strong>The complete HWF search path under a 6 × 6 search area. The black values represent the block weights calculated at altitude h0, while the blue values represent the weight scores calculated at altitude h1 .</p><p>A threshold for convenient visits, which is called the convenient threshold, is defined in advance. If the UAV passes through an intermediate block with a weight exceeding the convenient threshold along the HWF flight path from the current block to the next one, it will be guided to the center of this block for the convenient visit. This approach reduces the likelihood of unnecessary, repeated UAV flights among blocks. The value of the convenient threshold should be strategically defined to strike a balance. If it is set too high, the HWF variant could result in a UAV traversal path similar to that of the original HWF. Conversely, if it is set too low, the HWF variant could lead to excessive detours, which may diminish the HWF search efficiency.</p><p>We propose two convenient visit algorithms based on the HWF approach as follows. HWFR-S: A static and fixed convenient visit threshold is set. If there is a block along the HWF flight path that exceeds this convenient visit threshold, it will be visited by the UAV accordingly. In Figure <a href="#bookmark43">8</a>, the redpath represents the original HWF route, while the blue path represents the HWFR-S traversal route. HWFR-S sets a fixed convenient visit threshold of 0.7. Compared to the HWF algorithm, which guides the UAV from the block with a weight of 0.9 to the block with a weight of 0.85, HWFR-S reroutes to the intermediate block with a weight of 0.74 along the way. However, it does not choose to reroute to the block with a weight of 0.49 when the UAV flies from the block with a weight of 0.85 to the</p><p><a id="bookmark43"></a>block with a weight of 0.58, because the weight of 0.49 does not exceed the convenient visit threshold of 0.7.</p><p><img src="/media/202408//1724838587.4034371.png" /></p><p><strong>Figure 8. </strong>The original path (red line) of HWF and the modified path (blue line) of HWFR-S search algorithms.</p><p>HWFR-D: This algorithm proposes to use a fixed weight difference to calculate a dynamic convenient visit threshold for the next block. Its value is equal to the difference between the weight of the next block and the fixed weight difference, which means that the current visit threshold is dynamically adjusted by subtracting the weight difference from the next block’s weight value. If there is an intermediate block with a weight no less than the current convenient visit threshold, the UAV will take a detour to visit this intermediate block and then return to visit the original next block, instead of visiting the original next block directly.</p><p>In Figure <a href="#bookmark44">9</a>, the redpath represents the original HWF route, while the purple path represents the HWFR-D route when HWFR-D sets the fixed weight difference as 0.1. With the original HWF algorithm, the UAV first visits the block with a weight of 0.74, followed by the block with a weight of 0.58, and finally the block with a weight of 0.49. In contrast, when the UAV flies from the block with a weight of 0.74 to the block with a weight of 0.58, HWFR-D dynamically calculates its convenient visit threshold as 0.48, which is equal to 0.58 − 0.1. Then, HWFR-D chooses to conveniently visit the intermediate block with a weight of 0.49, because the weight of 0.49 is greater than the current convenient visit threshold of 0.48. In contrast, the intermediate block with the weight of 0.74 is not selected</p><p>for the convenient visit in the HWFR-D route when the UAV flies from the block with a weight of 0.9 to the block with a weight of 0.85, since the weight of 0.74 is less than the <a id="bookmark44"></a>current convenient visit threshold,i.e., 0.75 = 0.85 − 0.1.</p><p><img src="/media/202408//1724838587.4351711.png" /></p><p><strong>Figure 9. </strong>The original path (red line) of HWF and the modified path (purple dotted line) of HWFR-D search algorithms.</p><p>3.2.5. Flow of the Integrated YOLOv5 and HWF Framework</p><p>The flow of the integrated YOLOv5 and HWF framework, which tightly integrates the searching UAV equipped with the Jetson Nano and the cloud server at altitudes h0 and h1 in the search and rescue process, is illustrated in Figure <a href="#bookmark45">10.</a></p><p><img src="/media/202408//1724838587.51355.png" /></p><p><a id="bookmark45"></a><strong>Figure 10. </strong>Flowchart of the integrated YOLOv5 and HWF framework.</p><p>1. The user inputs the values of six parameters,i.e., the total search area size, the number of search blocks,and the features of the search target, including the types and colors of their clothing and pants, at the server side. At the same time, the UAV prepares for takeoff at the origin point.</p><p>2. The server calculates the center location of the search area and two flight heights, i.e., h0 and h1, for the searching UAV and sends this information to the Jetson Nano on the UAV through the mobile network communication.</p><p>3. The UAV flies to the specified center coordinates of the search area at altitude h0 and captures an image of the whole search area, which is then transmitted back to the server.</p><p>4. The server executes the weighting subroutine to calculate the first-level weights of all blocks within the total search area.</p><p>5. If the first-level weight of a particular block is greater than the search target threshold, the system proceeds to step 13; otherwise, it proceeds to step 6.</p><p>6. The system plans the second-level traversal path for the blocks at altitude h1, based on the first-level block weights, using the HWF algorithm. Then, the server transmits the planned path for the second layer to the Jetson Nano on the UAV.</p><p>7. The UAV flies to the center coordinates of the unvisited block with the highest first- level block weight at altitude h1, according to the planned path, and captures an image of the block. This block image is then transmitted back to the server.</p><p>8. If the UAV receives a command to finish the search, it proceeds to step 10; otherwise, it proceeds to step 9.</p><p>9. If all blocks with nonzero first-level weights have been visited by the UAV, the system proceeds to step 10; otherwise, it proceeds to step 7.</p><p>10. The UAV concludes its flight and returns to the starting point.</p><p>11. Whenever the server receives a block image transmitted by the UAV at step 7, it runs the weighting subroutine again to calculate the second-level block weight of the current search block.</p><p>12. If the second-level block weight is greater than the search target threshold, the system proceeds to step 13; otherwise, it returns to step 8.</p><p>13. The system outputs the coordinates of the detected target’s position along with its image to the user, which indicates that the search target has been found. The server then sends a command to the Jetson Nano on the UAV to finish the search mission.</p><p>3.2.6. Weighting Subroutine Flowchart</p><p>1. As shown in Figure <a href="#bookmark46">11</a>, the server first sets the initial values of the total weight value (W), human weights, and block weight as 0. It then executes YOLOv5 for human detection on UAV-captured images.</p><p>2. If a human body is detected, the server extracts the human body image with its bounding boxes and proceeds to use YOLOv5 for clothing recognition at step 3.</p><p>3. The server executes YOLOv5 for clothing recognition on the extracted human body image.</p><p>4. If the clothing is recognized and the recognized clothing type (Crct) matches the search clothing type (Csct), the total weight value (W) is incremented by the minimum value between the clothing accuracy (APca ) and the custom clothing fuzzy threshold of 0.3. Then, it proceeds to step 5. Otherwise, the system proceeds to step 6.</p><p>5. The server performs the KNN color recognition on the recognized clothing.</p><p>6. If the recognized clothing color (Crcc ) matches the search clothing color (Cscc ), the total weight value (W) is incremented by the minimum value between the KNN color percentage and the custom clothing color fuzzy threshold of 0.2. Then, it proceeds to step 7.</p><p>7. If the pants are recognized and the recognized pant type (Crpt) matches the search pant type (Cspt), the total weight value (W) is incremented by the minimum value between the pant accuracy (APpa ) and the custom pant fuzzy threshold of 0.3. The system proceeds to step 8; otherwise, it proceeds to step 9.</p><p>8. The server performs KNN color recognition on the recognized pants.</p><p>9. If the recognized pant color (Crpc ) matches the search pant color (Cspc ), the total weight value (W) is incremented by the minimum value between the KNN color percentage and the custom pant color fuzzy threshold of 0.2.</p><p>10. The weight score of a person (human_weight) is calculated by the weighted function of the human accuracy value (APha) and the total weight value (W), with the coefficient of 0.1 and 0.9, respectively. The maximum person weight value within a block is defined as the block_weight.</p><p><img src="/media/202408//1724838587.760424.png" /></p><p><a id="bookmark46"></a><strong>Figure 11. </strong>Flowchart of the weighting subroutine.</p><p>3.2.7. KNN Color Recognition Process</p><p>The flowchart of the KNN color recognition process is shown in Figure <a href="#bookmark47">12.</a> Its details <a id="bookmark47"></a>are listed below.</p><p><img src="/media/202408//1724838587.865717.png" /></p><p><strong>Figure 12. </strong>Flowchart of the KNN color recognition process.</p><p>1. After correctly identifying the types of clothing and pants, the system proceeds to extract clothing and pant images using the detected bounding box coordinates. Subsequently, the system applies noise reduction techniques to the image, facilitating the extraction of feature colors.</p><p>2. The captured clothing and pant images are converted from the RGB color space to the HSV color space. Subsequently, OpenCV [<a href="#bookmark48">37</a>] is used to generate a color gradient direction histogram for the image. From this histogram, the algorithm selects the interval with the highest proportion, obtaining a new set of HSV values, which serves as a representation of the image’s feature color.</p><p>3. The feature color representing the HSV color attributes is converted back to RGB color attributes.</p><p>4. The color distances between the image’s feature color and the RGB color table estab- lished in this study are computed. Subsequently, these distances are arranged in order, and k-nearest colors are chosen by a voting process. The color that receives the most votes is identified as the result of the KNN color recognition process.</p><p><strong>4. Simulation Results</strong></p><p>4.1. YOLOv5 Image Recognition Model</p><p>In this study, we utilize YOLOv5 for human body and clothing/pant recognition. These recognitions are executed on a cloud server. The human body database utilized for training in this study is sourced from VisDrone2019 [<a href="#bookmark49">38</a>], which is a dataset containing human-related images captured from the perspective of UAVs. It comprises a total of 7400 images, with 5600 images for training, 530 for validation, and 1270 for testing. The clothing/pant dataset is constructed by us and named the drone-clothing dataset, which contains a total of 5000 images, with 4000 images used for training, 500 for validation, and 500 for testing. It is also a dataset captured from the perspective of a UAV. The clothing and pants are categorized into four distinct types: short sleeves, long sleeves, short pants, and long pants. Both models are trained for 150,000 iterations on a server over the duration of one month.</p><p>The precision and mean average precision (mAP) values of the VisDrone2019 model are illustrated in Figures <a href="#bookmark50">13</a>and <a href="#bookmark51">14.</a> Precision measures whether the recognized target is indeed the intended target when it is detected, while mAP measures the confidence of the recognition when a target is detected and incorporates a trade-off between precision and recall. Figure <a href="#bookmark52">15</a>presents the original images and recognition results of the VisDrone2019 model. Because most original images in the VisDrone2019 model are shot at long distances, as shown in Figure <a href="#bookmark52">15</a>, which results in smaller and more complex objects, the recognized results are more likely to output incorrect results. Hence, the precision and mAP values of the training set in the VisDrone2019 model, shown in Figures <a href="#bookmark50">13</a>and <a href="#bookmark51">14</a>, are relatively stable after epoch 160 and reach 0.6773 and 0.5020 at epoch 460, respectively. Table <a href="#bookmark53">2</a> lists the values of the three classification metrics,i.e., the precision, recall, and mAP values, of the training set and the testing set in the VisDrone2019 model at epoch 460. These values for the training set are not notably higher than those of the testing set, which means that human body recognition using the VisDrone2019 model shows no significant overfitting behavior in this study. On the other hand, the precision and mean average precision (mAP) values of the drone-clothing model are illustrated in Figures <a href="#bookmark54">16</a> and <a href="#bookmark55">17.</a> Figure <a href="#bookmark56">18</a>presents the original images and recognition results of the drone-clothing model. As shown in Figure <a href="#bookmark56">18,</a> because most original images in the drone-clothing model are shot at distances closer to the human body than those in the VisDrone2019 model, the recognition results for human clothing are more precise. Hence, the precision and mAP values of the training set in the drone-clothing model, shown in Figures <a href="#bookmark54">16</a>and <a href="#bookmark55">17</a>, are relatively stable after epoch 40 and reach 0.9556 and 0.9540 at epoch 160, respectively. Consequently, the drone-clothing model converges faster and achieves higher classification metric values than the VisDrone2019 <a id="bookmark50"></a>model does.</p><p><img src="/media/202408//1724838588.120654.png" /></p><p><a id="bookmark51"></a><strong>Figure 13. </strong>Precision values (blue line) of the VisDrone2019 model.</p><p><img src="/media/202408//1724838588.2969458.png" /></p><p><strong>Figure 14. </strong>mAP values (blue line) of the VisDrone2019 model.</p><p><a id="bookmark52"></a><strong>Original Image Recognized Result</strong></p><p><img src="/media/202408//1724838588.487288.png" /><img src="/media/202408//1724838588.568131.png" /></p><p><img src="/media/202408//1724838588.790795.png" /><img src="/media/202408//1724838588.8008761.png" /><img src="/media/202408//1724838588.809143.png" /></p><p><img src="/media/202408//1724838588.887923.png" /></p><p><a id="bookmark53"></a><strong>Figure 15. </strong>Original images and recognition results of the VisDrone2019 model.</p><p><strong>Table 2. </strong>Values of classification metrics of the training set and the testing set in the Vis- Drone2019 model.</p><table><tr><td><p><strong>Classification Metric</strong></p></td><td><p><a id="bookmark54"></a><strong>Training Set</strong></p></td><td><p><strong>Testing Set</strong></p></td></tr><tr><td><p>Precision</p></td><td><p>0.6773</p></td><td><p>0.6330</p></td></tr><tr><td><p>Recall</p></td><td><p>0.4887</p></td><td><p>0.4074</p></td></tr><tr><td><p>mAP</p></td><td><p>0.5020</p></td><td><p>0.3970</p></td></tr></table><p><img src="/media/202408//1724838588.902959.png" /></p><p><strong>Figure 16. </strong>Precision values (blue line) of the drone-clothing model.</p><p><img src="/media/202408//1724838588.91387.png" /></p><p><strong>Figure 17. </strong>mAP values (blue line) of the drone-clothing model. <a id="bookmark55"></a><a id="bookmark56"></a><strong>Original Image Recognized Result</strong></p><p><img src="/media/202408//1724838588.92664.png" /><img src="/media/202408//1724838588.932109.png" /></p><p><img src="/media/202408//1724838588.95172.png" /><img src="/media/202408//1724838588.97911.png" /></p><p><img src="/media/202408//1724838588.985645.png" /><img src="/media/202408//1724838588.992918.png" /></p><p><strong>Figure 18. </strong>Original images and recognition results of the drone-clothing model.</p><p>4.2. Simulation Environment for Search Algorithms</p><p>In the simulation environment, Python is primarily used as the main programming language. The simulation area is partitioned into block sizes ranging from 3 × 3 to 10 × 10 and each block is assigned a weight. A total of 10,000 testing data are generated and each of them only gives a search target, which is randomly distributed within a specific block. We compare the proposed HWF, HWFR-S, and HWFR-D search algorithms with the sweep line and the spiral search algorithms in terms of four performance metrics,i.e., the average search pathlength, the average number of search blocks, the average search time, and the average search accuracy, required to locate the target location during the simulation of a UAV flight.</p><p>For each set of testing data, a single search target is randomly placed in a block within the n × n search area. Assume that the AP value of each search target in the captured image has different error ranges, depending on the UAV altitude. The features, i.e., the clothing type, pant type, clothing color, and pant color, of the search target are given different probabilities to change to different types and colors. The weighting subroutine calculates the weight value,i.e., human_weight, of each recognized person in every block based on the AP values of the four target features and determines whether the weight value of this person exceeds the search target weight threshold. The six simulation conditions are listed</p><p>below. The simulation parameters are shown in Table <a href="#bookmark57">3.</a> <a id="bookmark57"></a><strong>Table 3. </strong>Simulation parameters.</p><table><tr><td><p><strong>Parameter</strong></p></td><td><p><strong>Value</strong></p></td></tr><tr><td><p>Search area (m2 )</p></td><td><p>{(3n)2 | n = 3, 4, . . ., 10}</p></td></tr><tr><td><p><em>θ</em></p></td><td><p>40 degrees</p></td></tr><tr><td><p>Altitude hi (m)</p></td><td><p>√area 2tan<em>θ</em></p></td></tr><tr><td><p>The percentage of error and probability variation at altitude h0</p></td><td><p>10%, 20%, 30%</p></td></tr><tr><td><p>The percentage of error and probability variation at altitude h1</p></td><td><p>5%, 10%, 15%</p></td></tr><tr><td><p>UAV velocity</p></td><td><p>20 km/h</p></td></tr><tr><td><p>UAV hovering and image capture time for a block</p></td><td><p>5 s</p></td></tr><tr><td><p>Search target threshold</p></td><td><p>0.7</p></td></tr><tr><td><p>Convenient visit threshold of HWFR-S</p></td><td><p>0.4, 0.5, 0.6</p></td></tr><tr><td><p>Weight difference of HWFR-D</p></td><td><p>0.1, 0.2, 0.3</p></td></tr></table><p>1. The AP values for the search target’s human body, types of clothing and pants, and colors of clothing and pants are randomly distributed between 0.9 and 0.99. The search target is randomly assigned to a specific block. In contrast, the AP values of the human body and types of clothing and pants for the person that is not the search target are randomly distributed between 0.01 and 0.99. The AP values of their clothing and pant colors are randomly set between 0.1 and 0.9.</p><p>2. Each block contains one to four persons within it, with the probability of 70%. It has a 30% probability of having no person in one block, which means that the weight of this block is set to zero accordingly.</p><p>3. At a higher altitude h0, a larger error and probability variation of N% is applied to the given AP values and each feature of the person, respectively. Three different N% values,i.e., 10%, 20%, and 30%, are given to evaluate the performance metrics of these searching schemes.</p><p>4. At a lower altitude h1, a smaller error and probability variation of 0.5 N% is applied to the given AP values and each feature of the person, respectively. Hence, three different 0.5 N% values,i.e., 5%, 10%, and 15%, are set accordingly.</p><p>5. Because most available USB cameras support resolutions such as 640 × 480 or 1280 × 720, we assume that the side length of each block is limited to 3 m, and</p><p>the area of each block is 9 square meters, such that the USB camera can capture vivid images for the whole search area and each block. Therefore, the total search area for an n × n block is (3n)2 square meters.</p><p>6. As mentioned in Section <a href="#bookmark41">3.2.4,</a> HWFR-S instructs the UAV to visit the block that exceeds the static and fixed convenient visit threshold. If the value of the convenient visit threshold is too small, the UAV has a higher probability of rerouting to a block without the search target, which increases the search pathlength and search time accordingly. In contrast, if the value of the convenient visit threshold is too large, the UAV may lose the opportunity to reroute to the block with the search target. Hence, the convenient visit threshold of HWFR-S is given intermediate values between (0, 1) as 0.4, 0.5, and 0.6 in this simulation. HWFR-D is given a fixed weight difference to calculate a dynamic convenient visit threshold by subtracting the weight difference from the next block’s weight value. If there is an intermediate block with a weight no less than the current convenient visit threshold, the UAV will take a detour to visit this intermediate block. If the value of the weight difference is too large, which results in a smaller convenient visit threshold, the UAV suffers from a longer search path length and search time due to HWFR-S. Hence, the weight difference of HWFR-D is given lower values of 0.1, 0.2, and 0.3 in this simulation.</p><p>4.2.1. Average Search Path Length</p><p>As illustrated in Figure <a href="#bookmark58">19</a>, the spiral and sweep search algorithms do not prioritize the searching order of blocks that may potentially contain the target. Under varying errors and probability variations at altitudes h0 and h1, as shown in Figure <a href="#bookmark58">19</a>a–c, the average search pathlengths of these two algorithms with 10,000 testing data are nearly identical. Furthermore, as the block size increases, the search pathlengths of both algorithms grow linearly, resulting in the top two longest search pathlengths among the five search algo- rithms considered. It should be noted that these two algorithms have the same search pathlengths, as shown in Figure <a href="#bookmark58">19</a>a–c, regardless of the set of N% and 0.5 N% error and probability variations at altitudes h0 and h1 given in the simulation, because they follow their fixed block search patterns, which are not dependent on the N value.</p><p>As the HWF algorithm and its variants prioritize the searching order of blocks that may potentially contain the target, they exhibit significant reductions in their search path lengths compared to the spiral and sweep algorithms. As mentioned above, the HWF path planning algorithm performs a complete search over all blocks with nonzero weights and it may lead to a UAV search path that passes through some intermediate blocks several times, especially when none of the block weights surpass the search target threshold. Hence, HWF can lead to an increase in the search pathlength.</p><p>The HWFR-S algorithm, due to its fixed convenient visit threshold, becomes more effective at reducing its search pathlength when more block weights exceed this thresh- old. If the threshold is set lower, the search path length of HWFR-S becomes shorter. Hence, HWFR-S with the convenient visit threshold of 0.4, denoted as HWFR-S (0.4) in Figure <a href="#bookmark58">19</a>a–c, achieves shorter search pathlengths than HWFR-S (0.5) and HWFR-S (0.6). On the other hand, when the weight difference is larger, which results in a lower conve- nient visit threshold, the variant of HWFR-D achieves shorter search pathlengths. Hence, HWFR-D with the weight difference of 0.3, denoted as HWFR-D (0.3) in Figure <a href="#bookmark58">19</a>a–c, achieves shorter search pathlengths than HWFR-D (0.2) and HWFR-D (0.1). Hence, the HWFR-D algorithm has the capability to dynamically adjust the convenient visit threshold during the search process, making it the most successful algorithm to shorten the search pathlengths. Moreover, if a larger error and probability variation on the AP value is given in the simulation, the difference between the calculated first-/second-level weight value (based on the AP values of the four target features of each recognized person in every block) and its correct value becomes larger. Hence, the HWF algorithm and its variants have a higher probability of deploying the UAV to an incorrect block to search for the search target, which results in a longer search pathlength. As shown in Figure <a href="#bookmark58">19</a>a–c, the higher the N%</p><p>and 0.5 N% values are, the longer the search pathlengths of the HWF algorithm and all its HWFR-S and HWFR-D variants. It should be noted that the HWFR-D algorithm with a <a id="bookmark58"></a>larger error and probability variation outperforms HWFR-S and HWF significantly.</p><p><img src="/media/202408//1724838589.025952.png" /><img src="/media/202408//1724838589.031744.png" /></p><p>(<strong>a</strong>) (<strong>b</strong>)</p><p><img src="/media/202408//1724838589.052921.png" /></p><p>(<strong>c</strong>)</p><p><strong>Figure 19. </strong>Average search pathlength vs. search area size for (<strong>a</strong>) 10%, 5%; (<strong>b</strong>) 20%, 10%; (<strong>c</strong>) 30%, 15% error and probability variations at altitudes h0 and h1 .</p><p>4.2.2. Average Number of Search Blocks</p><p>As depicted in Figure <a href="#bookmark59">20</a>a–c, both the spiral and sweep search algorithms do not prioritize the search order of blocks that may potentially contain the target. With fewer than 10,000 testing data,the average number of search blocks required before finding the search target for these two algorithms is similar. Moreover, as the search area size increases, the number of search blocks of both algorithms increases linearly, causing them to display the highest number of search blocks among the five algorithms considered. It should be noted that these two algorithms have the same number of search blocks, as shown in Figure <a href="#bookmark59">20</a>a–c, regardless of the set of N% and 0.5 N% error and probability variations at altitudes h0 and h1 given in the simulation, because they follow their fixed block search patterns, which are not dependent on the N value.</p><p>The HWF, HWFR-S, and HWFR-D algorithms all share the characteristic of prioritizing the search order of blocks that is most likely to contain the search target. Since the first step in these algorithms is to select the block with the highest weight, the probability of finding the target at the first block is very high. This scenario results in the number of search blocks being equal to 1 for all three algorithms.</p><p>In the second scenario, when the HWF, HWFR-S, and HWFR-D algorithms encounter a situation wherein none of the block weights exceed the search target threshold during the search process, they will visit all blocks once and report that the search target cannot be found. In this case, the number of search blocks is equal to n × n for all three algorithms.</p><p><img src="/media/202408//1724838589.1114812.png" /><img src="/media/202408//1724838589.1676521.png" /></p><p><a id="bookmark59"></a>(<strong>a</strong>) (<strong>b</strong>)</p><p><img src="/media/202408//1724838589.230699.jpeg" /></p><p>(<strong>c</strong>)</p><p><strong>Figure 20. </strong>The average number of search blocks vs. search area size for (<strong>a</strong>) 10%, 5%; (<strong>b</strong>) 20%, 10%;</p><p>(<strong>c</strong>) 30%, 15% error and probability variations at altitudes h0 and h1 .</p><p>In the third scenario,the search target is found after the first step. HWFR-D exhibits an increase in the number of search blocks as its weight difference is larger. This means that HWFR-D with a larger weight difference will lower its convenient visit threshold and has a higher probability of rerouting to more intermediate blocks. In Figure <a href="#bookmark59">20</a>a–c, the weight difference is set to 0.3,i.e., HWFR-D (0.3) has a larger number of search blocks than HWFR-D (0.2) and HWFR-D (0.1). In contrast, when the fixed convenient visit threshold in HWFR-S is set lower, there is a slight increase in the number of search blocks for HWFR-S. Hence, HWFR-S (0.6) achieves the lowest number of search blocks in Figure <a href="#bookmark59">20</a>a–c. As mentioned above, if a larger error and probability variation on the AP value is given in the simulation, the difference between the calculated the first-/second-level weight value (based on the AP values of the four target features of each recognized person in every block) and its correct value becomes larger. Hence, the HWF algorithm and its variants have a higher probability of deploying the UAV to an incorrect block to search for the search target, which also results in a larger number of search blocks, as shown in Figure <a href="#bookmark59">20</a>a–c.</p><p>4.2.3. Average Search Time</p><p>The search time in this study is defined as the sum of the UAV flying time spent on the search path and the UAV hovering and image capture time for all visited blocks. In Figures <a href="#bookmark58">19</a> and <a href="#bookmark59">20</a>, the spiral and sweep algorithms exhibit nearly identical and linear growth in both the search path length and the number of search blocks. As a result, the calculated search times for these two algorithms are nearly identical, as shown in Figure <a href="#bookmark60">21</a>a–c.</p><p><img src="/media/202408//1724838589.4360478.png" /><img src="/media/202408//1724838589.442462.png" /></p><p><a id="bookmark60"></a>(<strong>a</strong>) (<strong>b</strong>)</p><p><img src="/media/202408//1724838589.454311.png" /></p><p>(<strong>c</strong>)</p><p><strong>Figure 21. </strong>Average search time vs. search area size for (<strong>a</strong>) 10%, 5%; (<strong>b</strong>) 20%, 10%; (<strong>c</strong>) 30%, 15% error and probability variations at altitudes h0 and h1 .</p><p>The HWF, HWFR-S, and HWFR-D algorithms exhibit a significant difference in their search times compared to the spiral and sweep algorithms. Under the parameters given in Table <a href="#bookmark57">3</a>, with a block size of 10 × 10, their search time is effectively reduced by over 80%, 75%, and 60%, as shown in Figure <a href="#bookmark60">21</a>a–c, respectively. In situations where the convenient visit threshold for HWFR-S and the weight difference for HWFR-D are not set appropriately, the search time for the HWF algorithm can be lower than in some HWFR-S and HWFR- D algorithms.</p><p>The HWFR-S algorithm exhibits its longest search time among the three fixed con- venient visit thresholds when the threshold is set to the highest value of 0.6. The search time for HWFR-S (0.6) is also higher than that of the original HWF algorithm. When the fixed convenient visit threshold of HWFR-S is lowered to 0.5, the average search time is not significantly different from that when the threshold is set to 0.4. The HWFR-D algorithm exhibits its longest search time when the weight difference is set to the highest value of 0.3, compared to both the HWF algorithm and HWFR-S with three different fixed convenient visit thresholds. However, when the HWFR-D weight difference is set to 0.2, it results in a shorter search time among the HWF algorithm and HWFR-S with three different fixed convenient visit thresholds. Furthermore, as the weight difference of HWFR-D decreases, the search time becomes shorter. When the weight difference is set to 0.1, the HWFR-D algorithm, as shown in Figure <a href="#bookmark60">21</a>a–c, achieves the shortest search time among all considered algorithms. As mentioned above, if a larger error and probability variation on the AP value is given in the simulation, the HWF algorithm and its variants have a higher probability of</p><p>deploying the UAV to an incorrect block to search for the search target, which also results in a longer search time, as shown in Figure <a href="#bookmark60">21</a>a–c.</p><p>4.2.4. Average Search Accuracy</p><p>When the search algorithm encounters a block having a recognized person with a weight greater than the target threshold, it concludes that the search target has been found. The search accuracy in this study is defined as the probability that the target found by the search algorithm is the same as the actual search target given in the testing data.</p><p>As shown in Figure <a href="#bookmark61">22</a>a–c, the spiral and sweep algorithms do not prioritize the searching order of blocks that may potentially contain the target. With (10%, 5%), (20%, 10%), and (30%, 15%) error and probability variations on features in the testing data at altitudes h0 and h1, respectively, it is observed that these two search algorithms, when applied to the 10 × 10 area size, achieve 50%, 40%, and 30% accuracy, respectively, in determining the actual search target after averaging over 10,000 testing data. The reason for this low accuracy is that these two algorithms often find a false target due to the high percentages of errors and probability variations when they encounter blocks with weights above the target threshold. Moreover, the HWF algorithm and its variants also suffer from lower search accuracy if a larger error and probability variation on the AP value is given in <a id="bookmark61"></a>the simulation, which is shown in Figure <a href="#bookmark61">22</a>a–c.</p><p><img src="/media/202408//1724838589.519113.png" /><img src="/media/202408//1724838589.560077.png" /></p><p>(<strong>a</strong>) (<strong>b</strong>)</p><p><img src="/media/202408//1724838589.592416.jpeg" /></p><p>(<strong>c</strong>)</p><p><strong>Figure 22. </strong>Average search accuracy vs. search area size for (<strong>a</strong>) 10%, 5%; (<strong>b</strong>) 20%, 10%; (<strong>c</strong>) 30%, 15% error and probability variations at altitudes h0 and h1 .</p><p>HWF, HWFR-S, and HWFR-D all share the characteristic of prioritizing the search order of blocks that is most likely to contain the target. Therefore, compared to search algorithms such as spiral and sweep, these algorithms achieve significantly higher accuracy.</p><p>Even with an area size of 10 × 10, the HWF-based algorithms maintain accuracy higher than 85%, 75%, and 65%, respectively, as shown in Figure <a href="#bookmark61">22</a>a–c. Furthermore, as observed from Figure <a href="#bookmark59">20</a>, different values of the convenient visit threshold and weight difference do not significantly affect the average accuracy of HWFR-S and HWFR-D after averaging over 10,000 testing data. Consequently, these HWFR-S and HWFR-D algorithms can reduce the search pathlength and decrease the search time, which in turn enhances the search efficiency.</p><p>In summary, the higher the error and probability variation applied to the AP values and each feature of the person is, the longer the search pathlength, the larger the number of search blocks, the longer the search time, and the lower the search accuracy of the HWF algorithm and all its HWFR-S and HWFR-D variants.</p><p><strong>5. System Implementation</strong></p><p>5.1. Software and Hardware</p><p>The searching UAV in this study adopts Pixhawk2.4.8 [<a href="#bookmark62">39</a>] as the core platform, en- hanced by the M8N high-precision GNSS/GPS positioning and navigation module, a GY-271M compass sensor module, a GY-US42V2 ultrasonic sensor range finder, 2212/13T brushless motors, a DEVO-RX1002 radio receiver, a DEVO10 remote controller, a PPM remote decoder, an electronic speed controller, a voltage indicator, an 8-inch nylon propeller, and a 915 MHz 100 mW servo-side transmitter. The communication protocol employed is 2.4 GHz (DSSS). Additionally, an Nvidia Jetson Nano [<a href="#bookmark36">34</a>] and USB camera are mounted on the UAV. The server environment operates on Windows 10, with hardware specifica- tions including an Intel® Core™ i7-8700 CPU, 16 GB of RAM, and a GeForce RTX 2080 TI GPU. In terms of software, Python serves as the primary programming language, while DroneKit-Python [<a href="#bookmark63">40</a>] is employed for UAV flight control, and the UAV platform is Mission Planner [<a href="#bookmark64">41,</a><a href="#bookmark65">42]</a>.</p><p>5.2. Screenshots of the Implemented System</p><p>As shown in Figure <a href="#bookmark66">23</a>, the user inputs the values of the total search area size, the number of search blocks, and the features of the search target, including the types and colors of the clothing and pants, at the server side. The search target to be found is wearing a black long-sleeved shirt and black pants. After the user inputs the search information, the server computes the UAV flight altitudes h0 and h1 and transmits them to the Jetson Nano on the searching UAV to dispatch the UAV to fly to the center of search area at altitude h0 . The Jetson Nano on the UAV captures images at altitude h0 and transmits them back to the server. The server first partitions the search area into 9 blocks, executes the weighting subroutine to calculate the first-level weights of all blocks, and calculates the HWF block traversal order for the UAV, which is block [35.06] → block [9.0] → block [3.39] in this example, as depicted in Figure <a href="#bookmark67">24.</a> Subsequently, the server transmits this block traversal order to the Jetson Nano.</p><p>After receiving the block traversal order, the Jetson Nano commands the UAV to fly and descend to the center of the block with the highest weight,i.e., block [35.06],at altitude h1 . The UAV further captures images of this block and transmits them back to the server. Using the weighting subroutine, the server calculates the weight value of each recognized person in this block and determines whether the second-level weight value is larger than the search target threshold. In Figure <a href="#bookmark68">25</a>, since the person is wearing a red long-sleeved shirt and black pants, the computed second-level weight of this person is 0.7534,which is lower than the search target threshold of 0.8. Hence, the UAV continues the HWF search in the next block with a first-level weight of 9.0.</p><p>In Figure <a href="#bookmark69">26</a>, because the recognized person in block [9.0] is wearing a black long- sleeved shirt and black pants, the calculated second-level weight value for this person is 0.9052, which is greater than the target threshold of 0.8. Hence, the algorithm determines that the search target has been found. The Jetson Nano on the UAV then transmits photos</p><p>of the person found in this block, along with the location coordinates, back to the server to <a id="bookmark66"></a>notify the user, which concludes this search mission.</p><p><img src="/media/202408//1724838589.683192.png" /></p><p><a id="bookmark67"></a><strong>Figure 23. </strong>User inputs search information at server.</p><p><img src="/media/202408//1724838589.7103748.png" /></p><p><strong>Figure 24. </strong>The HWF block traversal order (green line) in search area. The red lines and the blue values indicate the boundaries and the first-level weights of all blocks respectively.</p><p><img src="/media/202408//1724838589.749782.png" /></p><p><a id="bookmark68"></a><a id="bookmark69"></a><strong>Figure 25. </strong>Recognized results for the person in the block with weight 35.06 at altitude h1 .</p><p><img src="/media/202408//1724838589.938695.png" /></p><p><strong>Figure 26. </strong>Recognized results for the person in the block with weight 9.0 at altitude h1 .</p><p><a id="bookmark70"></a>5.3. Limitations of the Proposed System</p><p>There are important prerequisites for the integrated YOLOv5 and hierarchical HWF path planning framework proposed in this study, which are as follows.</p><p>1. The ideal search area must approximate a rectangle. If it is a concave or convex poly- gon or any irregular shape, the input range for HWF must be the smallest bounding rectangle that encompasses the entire search area. This would expand the search area, potentially including many non-search areas, leading to longer search paths and times, as well as increased power consumption for the UAV.</p><p>2. The ideal altitude for the search area should be consistent across a single horizontal plane. This ensures that when the UAV captures the image of the entire search area at altitude h0, the distances between the UAV and the center points of different blocks are similar. Hence, the relationship among the block weight values obtained from the human body and clothing recognition at the first level will closely approximate those of the real search target. Further, using HWF path planning to visit blocks at the second level with the highest block weight value first and subsequently recognizing the results at the second level will yield more accurate outcomes. Conversely, if the altitudes of</p><p>the center points of different blocks are not on the same horizontal plane, the UAV will be closer to blocks at higher altitudes. This results in clearer, more magnified images of human bodies, leading to better recognition results. Consequently, a block with higher block weight values might be prioritized in the HWF path planning algorithm. If the search target is not within this block, it could result in longer search paths and times.</p><p>3. Since the UAV captures images at altitude h0 in the first level, it must cover the entire area. Due to the limited resolution of the camera, the search area cannot be too expansive. Otherwise, the captured images of human bodies would appear smaller and blurrier, leading to poorer recognition results and subsequently affecting the accuracy of HWF path planning.</p><p>5.4. Performance Comparison of YOLOv5 and YOLOv8</p><p>YOLOv5 was released in June 2020. The YOLOv5 model has Darknet 53 as its backbone and its design focuses on enhancing the detection of objects at different scales, which improves the performance on objects of varying sizes. YOLOv8 is the latest version of the YOLO family, developed by Ultralytics, who also created the YOLOv5 model. It introduces numerous architectural changes over YOLOv5. Unlike YOLOv5, YOLOv8 is an anchor-free model, meaning that it directly predicts the center of an object instead of the offset from a known anchor box. The study conducted by [<a href="#bookmark71">43</a>] involved a comparative analysis of the performance of YOLOv5 and YOLOv8 in aerial human detection using unmanned aerial vehicles. The research utilized a pedestrian dataset obtained from Roboflow, consisting of 828 aerial images for model training and 233 images for validation. The experimental results revealed that the YOLOv8 model exhibited higher precision and F1-scores compared to the YOLOv5 model, with differences of 2.82% and 0.98%, respectively. However, in terms of recall performance, YOLOv5 surpasses the YOLOv8 model by 0.54%. In [<a href="#bookmark72">44</a>], a comparison was made between various versions of YOLO, specifically YOLOv5 toYOLOv8, based on their mean average precision (mAP) scores. The study involved the training of models using 2415 images, and 303 images were allocated for both validation and testing purposes. The images were categorized into five classes. The research findings indicated that YOLOv5 achieved the highest average detection accuracy in terms of the mAP metric, whereas YOLOv8 performed the poorest among the four versions. Considering these results collectively, it is evident that the choice of dataset significantly influences the performance of different YOLO versions. Therefore, for future work, we are contemplating the utilization</p><p>of the updated YOLOv8 to train our human detection model.</p><p><strong>6. Conclusions</strong></p><p>In this study, to reduce the search time and increase the search accuracy of search and rescue operations, an integrated YOLOv5 and HWF framework has been proposed. It combines the YOLOv5 model to automatically recognize the search target in realtime and the hierarchical HWF path planning algorithm to dispatch a UAV to capture images of the search target at different altitudes. Two improved search algorithms, HWFR-S and HWFR-D, which incorporate the concepts of the convenient visit threshold and the weight difference, respectively, have been further proposed to resolve the issue of the lengthy and redundant flight paths of HWF. YOLOv5 has been trained by the VisDrone2019 dataset and the drone-clothing dataset for human body and clothing/pant recognition, respectively. The results show that the drone-clothing model converges faster and achieves higher classification metric values than the VisDrone2019 model does. According to the simulation results, the HWF, HWFR-S, and HWFR-D search algorithms proposed in this study not only effectively reduce the length of the UAV’s search path and the number of search blocks but also significantly decrease the search time required for the UAV to locate the search target, with a much higher search accuracy than the two traditional search algorithms. Moreover, this integrated YOLOv5 and HWF framework has been implemented and tested</p><p>in a real scenario; it has been shown to reduce the UAV’s power consumption and enhance the efficiency of search and rescue operations.</p><p>In the future, we will address the issues and limitations mentioned in Section <a href="#bookmark70">5.3</a>, for example, the irregular shapes of search and rescue areas, the different altitudes of areas, and areas that exceed the resolution range of the UAV camera. Additionally, to reduce search and rescue times, the possibility of utilizing multiple UAVs in collaborative search and rescue operations will be considered. Further, we will evaluate the utilization of the updated YOLOv8 to train our human detection model and compare its performance with that of YOLOv5 as another future task. By refining the integrated YOLOv5 and hierarchical HWF path planning system proposed in this study, we aim to develop a more versatile UAV search and rescue system in the future.</p><p><strong>Author Contributions: </strong>Conceptualization,I.-C.C.; methodology, I.-C.C., H.-F.C., Y.-W.C., M.-T.H., W.-F.W. and D.-Y.Y.; software, H.-F.C., Y.-W.C., M.-T.H., W.-F.W. and D.-Y.Y.; validation,I.-C.C. and C.-E.Y.; formal analysis,I.-C.C.; resources, C.-E.Y.; data curation, H.-F.C., Y.-W.C., M.-T.H., W.-F.W. and D.-Y.Y.; writing—original draft preparation, Y.-H.H., I.-C.C. and C.-E.Y.; writing—review and editing, Y.-H.H., I.-C.C. and C.-E.Y.; supervision,I.-C.C. and C.-E.Y.; project administration,I.-C.C. and C.-E.Y.; funding acquisition,I.-C.C. All authors have read and agreed to the published version of the manuscript.</p><p><strong>Funding: </strong>This research was funded by the National Science and Technology Council, Taiwan, grant number NSTC 112-2221-E-018-009. The APC was funded by NSTC 112-2221-E-018-009.</p><p><strong>Data Availability Statement: </strong>Data are contained within the article. <strong>Conflicts of Interest: </strong>The authors declare no conflicts of interest.</p><p><strong>References</strong></p><p>1. Sahingoz, O.K. Networking models in flying ad-hoc networks (FANETs): Concepts and Challenges. J. Intell. Robot. Syst. <strong>2014</strong>,</p><p><a id="bookmark1"></a><a id="bookmark2"></a>74, 513–527. <a href="https://doi.org/10.1007/s10846-013-9959-7">[CrossRef]</a></p><p>2. Menouar, H.; Guvenc,I.; Akkaya, K.; Uluagac, A.S.; Kadri, A.; Tuncer, A. UAV-enabled intelligent transportation systems for the <a id="bookmark3"></a>smart city: Applications and challenges. IEEE Commun. Mag. <strong>2017</strong>, 55, 22–28. <a href="https://doi.org/10.1109/MCOM.2017.1600238CM">[CrossRef]</a></p><p>3. Aasen, H. UAV spectroscopy: Current sensors, processing techniques and theoretical concepts for data interpretation. In Proceed- <a id="bookmark4"></a>ings of the IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 8809–8812.</p><p>4. Ezequiel, C.A.F.; Cua, M.; Libatique, N.C.; Tangonan, G.L.; Alampay, R.; Labuguen, R.T.; Favila, C.M.; Honrado,J.L.E.; Canos, V.; Devaney, C.; et al. UAV aerial imaging applications for post-disaster assessment, environmental management and infrastructure development. In Proceedings of the International Conference on Unmanned Aircraft Systems, Orlando, FL, USA, 27–30 May 2017; <a id="bookmark5"></a>pp. 274–283.</p><p>5. Zhang, Y.; Li, S.; Wang, S.; Wang, X.; Duan, H. Distributed bearing-based formation maneuver control of fixed-wing UAVs by <a id="bookmark6"></a>finite-time orientation estimation. Aerosp. Sci. Technol. <strong>2023</strong>, 136, 108241. <a href="https://doi.org/10.1016/j.ast.2023.108241">[CrossRef]</a></p><p>6. Zheng,Q.; Zhao, P.; Li, Y.; Wang, H.; Yang, Y. Spectrum interference-based two-level data augmentation method in deep learning <a id="bookmark7"></a>for automatic modulation classification. Neural Comput. Applic. <strong>2021</strong>, 33, 7723–7745. <a href="https://doi.org/10.1007/s00521-020-05514-1">[CrossRef]</a></p><p>7. Mao, Y.; Sun, R.; Wang,J.; Cheng,Q.; Kiong, L.C.; Ochieng, W.Y. New time-differenced carrier phase approach to GNSS/INS <a id="bookmark8"></a>integration. GPS Solut. <strong>2022</strong>, 26, 122. <a href="https://doi.org/10.1007/s10291-022-01314-3">[CrossRef]</a></p><p>8. Zhang, X.; Pan, W.; Scattolini, R.; Yu, S.; Xu, X. Robust tube-based model predictive control with Koopman operators. Automatica <a id="bookmark9"></a><strong>2022</strong>, 137, 110114. <a href="https://doi.org/10.1016/j.automatica.2021.110114">[CrossRef]</a></p><p>9. Narayanan, S.S.K.S.; Tellez-Castro, D.; Sutavani, S.; Vaidya, U. SE(3) (Koopman-MPC: Data-driven learning and control of <a id="bookmark10"></a>quadrotor UAVs. IFAC-PapersOnLine <strong>2023</strong>, 56, 607–612. <a href="https://doi.org/10.1016/j.ifacol.2023.12.091">[CrossRef]</a></p><p>10. Cao, B.; Zhang,W.; Wang, X.; Zhao,J.; Gu,Y.; Zhang,Y. A memetic algorithm based on two_Arch2 for multi-depot heterogeneous- <a id="bookmark11"></a>vehicle capacitated arc routing problem. Swarm Evol. Comput. <strong>2021</strong>, 63, 100864. <a href="https://doi.org/10.1016/j.swevo.2021.100864">[CrossRef]</a></p><p>11. Erdelj, M.; Natalizio, E. UAV-assisted disaster management: Applications and open issues. In Proceedings of the International Conference on Computing, Networking and Communications, Kauai, HI, USA, 15–18 February 2016; pp. 1–5.</p><p>12. Mukherjee, A.; De, D.; Dey, N.; Crespo, R.G.; Herrera-Viedma, E. DisastDrone: A Disaster Aware Consumer Internet of Drone <a id="bookmark12"></a>Things System in Ultra-Low Latent 6G Network. IEEE Trans. Consum. Electron. <strong>2023</strong>, 69, 38–48. <a href="https://doi.org/10.1109/TCE.2022.3214568">[CrossRef]</a></p><p>13. Pasandideh, F.; daCosta,J.P.J.; Kunst, R.; Islam,N.; Hardjawana,W.; Pignatonde Freitas, E. A Review of Flying Ad Hoc Networks: <a id="bookmark13"></a>Key Characteristics, Applications, and Wireless Technologies. Remote Sens. <strong>2022</strong>, 14, 4459. <a href="https://doi.org/10.3390/rs14184459">[CrossRef]</a></p><p>14. Majeed, A.; Hwang, S.O. A Multi-Objective Coverage Path Planning Algorithm for UAVs to Cover Spatially Distributed Regions in Urban Environments. Aerospace <strong>2021</strong>, 8, 343. <a href="https://doi.org/10.3390/aerospace8110343">[CrossRef]</a></p><p><a id="bookmark19"></a>15. Das, L.B.; Das, L.B.; Lijiya, A.; Jagadanand, G.; Aadith, A.; Gautham, S.; Mohan, V.; Reuben, S.; George, G. Human Target Search and Detection using Autonomous UAV and Deep Learning. In Proceedings of the IEEE International Conference on Industry 4.0, <a id="bookmark14"></a>Artificial Intelligence, and Communications Technology (IAICT), Bali, Indonesia, 7–8 July 2020; pp. 55–61. <a href="https://doi.org/10.1109/IAICT50021.2020.9172031">[CrossRef]</a></p><p>16. Bandeira, T.W.; Coutinho, W.P.; Brito, A.V.; Subramanian, A. Analysis of Path Planning Algorithms Based on Travelling Salesman Problem Embedded in UAVs. In Proceedings of the Brazilian Symposium on Computing Systems Engineering (SBESC),Fortaleza, <a id="bookmark15"></a>Porto Alegre, Brazil, 3–6 November 2015; pp. 70–75. <a href="https://doi.org/10.1109/SBESC.2015.20">[CrossRef]</a></p><p>17. Jain, A.; Ramaprasad, R.; Narang, P.; Mandal, M.; Chamola, V.; Yu, F.R.; Guizan, M. AI-Enabled Object Detection in UAVs: <a id="bookmark16"></a>Challenges,Design Choices, and Research Directions. IEEE Netw. <strong>2021</strong>, 35, 129–135. <a href="https://doi.org/10.1109/MNET.011.2000643">[CrossRef]</a></p><p>18. Yu, X.; Jin, S.; Shi, D.; Li, L.; Kang, Y.; Zou, J. Balanced Multi-Region Coverage Path Planning for Unmanned Aerial Vehi- cles. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics (SMC), Toronto, ON, Canada, 11–14 October 2020; pp. 3499–3506.</p><p>19. Yaguchi, Y.; Tomeba, T. Region Coverage Flight Path Planning Using Multiple UAVs to Monitor the Huge Areas. In Proceedings of the IEEE International Conference on Unmanned Aircraft Systems (ICUAS), Athens, Greece, 15–18 June 2021; pp. 1677–1682.</p><p>20. Kurdi, H.A.; Aloboud, E.; Alalwan, M.; Alhassan, S.; Alotaibi, E.; Bautista, G.; How, J.P. Autonomous Task Allocation for Multi-UAV Systems Based on the Locust Elastic Behavior. Appl. Soft Comput. <strong>2018</strong>, 71, 110–126. <a href="https://doi.org/10.1016/j.asoc.2018.06.006">[CrossRef]</a></p><p>21. Alotaibi, E.T.; Alqefari, S.S.; Koubaa, A. LSAR-Multi-UAV Collaboration for Search and Rescue Missions. IEEE Access <strong>2019</strong>, <a id="bookmark17"></a><a id="bookmark21"></a>7, 55817–55832. <a href="https://doi.org/10.1109/ACCESS.2019.2912306">[CrossRef]</a></p><p>22. Cabreira, T.; Brisolara, L.; Ferreira, P.R., Jr. Survey on Coverage Path Planning with Unmanned Aerial Vehicles. Drones <strong>2019</strong>, 3, 4. <a id="bookmark23"></a><a href="https://doi.org/10.3390/drones3010004">[CrossRef]</a></p><p>23. Jünger, M.; Reinelt, G.; Rinaldi, G. The Traveling Salesman Problem. In Handbooks in Operations Research and Management Science; <a id="bookmark24"></a>Elsevier B.V.: Amsterdam, The Netherlands, 1995; Volume 7,pp. 225–330.</p><p>24. Ali, M.; Md Rashid, N.K.A.; Mustafah, Y.M. Performance Comparison between RGB and HSV Color Segmentations for Road <a id="bookmark25"></a>Signs Detection. Appl. Mech. Mater. <strong>2013</strong>, 393, 550–555. <a href="https://doi.org/10.4028/www.scientific.net/amm.393.550">[CrossRef]</a></p><p>25. Haritha, D.; Bhagavathi, C. Distance Measures in RGB and HSV Color Spaces. In Proceedings of the 20th International Conference <a id="bookmark26"></a>on Computers and Their Applications (CATA 2005), New Orleans, LA, USA, 16–18 March 2005.</p><p>26. Pooja, K.S.; Shreya, R.N.; Lakshmi, M.S.; Yashika, B.C.; Rekha, B.N. Color Recognition using K-Nearest Neighbors Machine <a id="bookmark27"></a>Learning Classification Algorithm Trained with Color Histogram Features. Int. Res. J. Eng. Technol. (IRJET) <strong>2021</strong>, 8, 1935–1936.</p><p>27. Pradeep, A.G.; Gnanapriya, M. Novel Contrast Enhancement Algorithm Using HSV Color Space. Int. J. Innov. Technol. Res. <strong>2016</strong>, 4, 5073–5074.</p><p><a id="bookmark28"></a>28. Krishna, S.L.; Chaitanya, G.S.R.; Reddy, A.S.H.; Naidu, A.M.; Poorna, S.S.; Anuraj, K. Autonomous Human Detection System Mounted on a Drone. In Proceedings of the 2019 International Conference on Wireless Communications Signal Processing and <a id="bookmark29"></a>Networking (WiSPNET), Chennai, India, 21–23 March 2019; pp. 335–338. <a href="https://doi.org/10.1109/WiSPNET45539.2019.9032876">[CrossRef]</a></p><p>29. Mliki, H.; Bouhlel, F.; Hammami, H. Human activity recognition from UAV-captured video sequences. Pattern Recognit. <strong>2020</strong>, <a id="bookmark30"></a>100, 107140. <a href="https://doi.org/10.1016/j.patcog.2019.107140">[CrossRef]</a></p><p>30. Safadinho, D.; Ramos,J.; Ribeiro, R.; Filipe, V.; Barroso, J.; Pereira, A. UAV Landing Using Computer Vision Techniques for <a id="bookmark31"></a>Human Detection. Sensors <strong>2020</strong>, 20, 613. <a href="https://doi.org/10.3390/s20030613">[CrossRef</a>] [<a href="https://www.ncbi.nlm.nih.gov/pubmed/31979142">PubMed]</a></p><p>31. Lygouras, E.; Santavas, N.; Taitzoglou, A.; Tarchanidis, K.; Mitropoulos, A.; Gasteratos, A. Unsupervised Human Detection with <a id="bookmark32"></a>an Embedded Vision System on a Fully Autonomous UAV for Search and Rescue Operations. Sensors <strong>2019</strong>, 19, 3542. <a href="https://doi.org/10.3390/s19163542">[CrossRef]</a></p><p>32. Do, M.-T.; Ha, M.-H.; Nguyen, D.-C.; Thai, K.; Ba,Q.-H.D. Human Detection Based Yolo Backbones-Transformer in UAVs. In Proceedings of the International Conference on System Science and Engineering (ICSSE), Ho Chi Minh, Vietnam, 27–28 July 2023; <a id="bookmark33"></a>pp. 576–580. <a href="https://doi.org/10.1109/ICSSE58758.2023.10227141">[CrossRef]</a></p><p>33. Wijesundara, D.; Gunawardena, L.; Premachandra, C. Human Recognition from High-altitude UAV Camera Images by AI based Body Region Detection. In Proceedings of the Joint 12th International Conference on Soft Computing and Intelligent Systems and 23rd International Symposium on Advanced Intelligent Systems (SCIS & ISIS), Ise, Japan, 29 November—2 December 2022; <a id="bookmark36"></a>pp. 1–4. <a href="https://doi.org/10.1109/SCISISIS55246.2022.10002039">[CrossRef]</a></p><p>34. Jetson Nano Developer Kit | NVIDIA. Available online: <a href="https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-nano-developer-kit/">https://www.nvidia.com/en-us/autonomous-machines/embedded-</a> <a id="bookmark37"></a><a href="https://www.nvidia.com/en-us/autonomous-machines/embedded-systems/jetson-nano-developer-kit/">systems/jetson-nano-developer-kit/</a> (accessed on 1 December 2023).</p><p>35. Itkin, M.; Kim, M.; Park, Y. Development of Cloud-Based UAV Monitoring and Management System. Sensors <strong>2016</strong>, 16, 1913. <a id="bookmark39"></a><a href="https://doi.org/10.3390/s16111913">[CrossRef]</a></p><p>36. Geng, X.; Chen, Z.; Yang, W.; Shi, D.; Zhao, K. Solving the Traveling Salesman Problem Based on an Adaptive Simulated <a id="bookmark48"></a>Annealing Algorithm with Greedy Search. Appl. Soft Comput. <strong>2011</strong>, 11, 3680–3689. <a href="https://doi.org/10.1016/j.asoc.2011.01.039">[CrossRef]</a></p><p><a id="bookmark49"></a>37. OpenCV—Open Computer Vision Library. Available online: <a href="https://opencv.org/">https://opencv.org/</a>(accessed on 1 December 2023).</p><p>38. VisDrone-Dataset-github. Available online: <a href="https://github.com/VisDrone/VisDrone-Dataset">https://github.com/VisDrone/VisDrone-Dataset</a> (accessed on 1 December 2023).</p><p><a id="bookmark62"></a><a id="bookmark63"></a>39. Pixhawk. Available online: <a href="https://pixhawk.org/">https://pixhawk.org/</a> (accessed on 1 December 2023).</p><p>40. Welcome to DroneKit-Python’s Documentation. Available online: <a href="https://dronekit-python.readthedocs.io/en/latest/">https://dronekit-python.readthedocs.io/en/latest/</a> (accessed <a id="bookmark64"></a>on 1 December 2023).</p><p>41. Mission Planner Home—Mission Planner Documentation (ardupilot.org). Available online: <a href="https://ardupilot.org/planner/">https://ardupilot.org/planner/</a> (accessed on 1 December 2023).</p><p><a id="bookmark65"></a>42. Suparnunt, C.; Boonvongsobhon, C.; Eounes Baig, F.; Leelalerthpat, P.; Hematulin, W.; Jarawan, T.; Kamsing, P. Practical Parallel of Autonomous Unmanned Aerial Vehicle by Mission Planner. In Proceedings of the IEEE International Geoscience and Remote <a id="bookmark71"></a>Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 7831–7834. <a href="https://doi.org/10.1109/IGARSS46834.2022.9884577">[CrossRef]</a></p><p>43. Sary, I.P.; Andromeda, S.; Armin, E.U. Performance Comparison of YOLOv5 and YOLOv8 Architectures in Human Detection <a id="bookmark72"></a>using Aerial Images. Ultim. Comput. J. Sist. Komputer. <strong>2023</strong>, 15, 8–13. <a href="https://doi.org/10.31937/sk.v15i1.3204">[CrossRef]</a></p><p>44. Gašparovi´c, B.; Mauša, G.; Rukavina, J.; Lerga, J. Evaluating YOLOV5, YOLOV6, YOLOV7, and YOLOV8 in Underwater Environment: Is There Real Improvement? In Proceedings of the 8th International Conference on Smart and Sustainable Technologies (SpliTech), Split/Bol, Croatia, 20–23 June 2023; pp. 1–4. <a href="https://doi.org/10.23919/SpliTech58164.2023.10193505">[CrossRef]</a></p><p><strong>Disclaimer/Publisher’s Note: </strong>The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.</p>
刘世财
2024年8月28日 17:49
转发文档
收藏文档
上一篇
下一篇
手机扫码
复制链接
手机扫一扫转发分享
复制链接
Markdown文件
HTML文件
PDF文档(打印)
分享
链接
类型
密码
更新密码