Please use this identifier to cite or link to this item:
https://ah.lib.nccu.edu.tw/handle/140.119/135538
題名: | Real-Time Vision-Based River Detection and Lateral Shot Following for Autonomous UAVs | 作者: | 劉吉軒 Jyi-Shane Liu Huang, Yenting Lee, Gongyi Soong, Rutai |
貢獻者: | 資科系 | 關鍵詞: | Rivers ; Image segmentation ; Unmanned aerial vehicles ; Navigation ; Inspection ; Task analysis ; Image edge detection | 日期: | 九月-2020 | 上傳時間: | 4-六月-2021 | 摘要: | Most existing autonomous UAV inspection tasks focus on environment surroundings and facilities. The UAV often navigates above the inspected target and conducts inspection with the camera aiming downward on the target. However, in some scenarios, it is risky to allow UAVs to navigate above the inspected target. For example, when patrolling a river, the UAV may risk falling into the river. Similar risks also exist for scenarios such as railways and power lines. This research proposes a lateral shot following approach for UAVs to follow the river laterally while collecting image data with a front view camera. The proposed approach has been evaluated with different segments of river in real world environments. The experiments include two types of following method and two types of viewpoint to suit different task needs. Results show that our deep neural network can extract the river masks in real-time with high accuracy. With adaptive steering adjustments, the UAV can achieve accurate and robust following when handling geographical change of river segments. Performance comparison between human operators and our developed autonomous system shows that better following accuracy and consistency can be achieved by our autonomous system. | 關聯: | Proceedings of the 2020 IEEE International Conference on Real-time Computing and Robotics, IEEE | 資料類型: | conference | DOI: | https://doi.org/10.1109/RCAR49640.2020.9303263 |
Appears in Collections: | 會議論文 |
Show full item record
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.