Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 1 de 1
Filtrar
Mais filtros

Base de dados
Ano de publicação
Tipo de documento
Intervalo de ano de publicação
1.
Sensors (Basel) ; 23(16)2023 Aug 21.
Artigo em Inglês | MEDLINE | ID: mdl-37631840

RESUMO

Unmanned vehicles frequently encounter the challenge of navigating through complex mountainous terrains, which are characterized by numerous unknown continuous curves. Drones, with their wide field of view and ability to vertically displace, offer a potential solution to compensate for the limited field of view of ground vehicles. However, the conventional approach of path extraction solely provides pixel-level positional information. Consequently, when drones guide ground unmanned vehicles using visual cues, the road fitting accuracy is compromised, resulting in reduced speed. Addressing these limitations with existing methods has proven to be a formidable task. In this study, we propose an innovative approach for guiding the visual movement of unmanned ground vehicles using an air-ground collaborative vectorized curved road representation and trajectory planning method. Our method offers several advantages over traditional road fitting techniques. Firstly, it incorporates a road star points ordering method based on the K-Means clustering algorithm, which simplifies the complex process of road fitting. Additionally, we introduce a road vectorization model based on the piecewise GA-Bézier algorithm, enabling the identification of the optimal frame from the initial frame to the current frame in the video stream. This significantly improves the road fitting effect (EV) and reduces the model running time (T-model). Furthermore, we employ smooth trajectory planning along the "route-plane" to maximize speed at turning points, thereby minimizing travel time (T-travel). To validate the efficiency and accuracy of our proposed method, we conducted extensive simulation experiments and performed actual comparison experiments. The results demonstrate the superior performance of our approach in terms of both efficiency and accuracy.

SELEÇÃO DE REFERÊNCIAS
Detalhe da pesquisa