Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 14 de 14
Filtrar
1.
Sensors (Basel) ; 24(16)2024 Aug 22.
Artículo en Inglés | MEDLINE | ID: mdl-39205117

RESUMEN

3D object-detection based on LiDAR point clouds can help driverless vehicles detect obstacles. However, the existing point-cloud-based object-detection methods are generally ineffective in detecting small objects such as pedestrians and cyclists. Therefore, a small-object-detection algorithm based on clustering is proposed. Firstly, a new segmented ground-point clouds segmentation algorithm is proposed, which filters out the object point clouds according to the heuristic rules and realizes the ground segmentation by multi-region plane-fitting. Then, the small-object point cloud is clustered using an improved DBSCAN clustering algorithm. The K-means++ algorithm for pre-clustering is used, the neighborhood radius is adaptively adjusted according to the distance, and the core point search method of the original algorithm is improved. Finally, the detection of small objects is completed using the directional wraparound box model. After extensive experiments, it was shown that the precision and recall of our proposed ground-segmentation algorithm reached 91.86% and 92.70%, respectively, and the improved DBSCAN clustering algorithm improved the recall of pedestrians and cyclists by 15.89% and 9.50%, respectively. In addition, visualization experiments confirmed that our proposed small-object-detection algorithm based on the point-cloud clustering method can realize the accurate detection of small objects.

2.
Sensors (Basel) ; 23(2)2023 Jan 05.
Artículo en Inglés | MEDLINE | ID: mdl-36679414

RESUMEN

In the near future, autonomous vehicles with full self-driving features will populate our public roads. However, fully autonomous cars will require robust perception systems to safely navigate the environment, which includes cameras, RADAR devices, and Light Detection and Ranging (LiDAR) sensors. LiDAR is currently a key sensor for the future of autonomous driving since it can read the vehicle's vicinity and provide a real-time 3D visualization of the surroundings through a point cloud representation. These features can assist the autonomous vehicle in several tasks, such as object identification and obstacle avoidance, accurate speed and distance measurements, road navigation, and more. However, it is crucial to detect the ground plane and road limits to safely navigate the environment, which requires extracting information from the point cloud to accurately detect common road boundaries. This article presents a survey of existing methods used to detect and extract ground points from LiDAR point clouds. It summarizes the already extensive literature and proposes a comprehensive taxonomy to help understand the current ground segmentation methods that can be used in automotive LiDAR sensors.


Asunto(s)
Conducción de Automóvil , Automóviles , Vehículos Autónomos , Espectrometría Raman
3.
Sensors (Basel) ; 23(7)2023 Mar 26.
Artículo en Inglés | MEDLINE | ID: mdl-37050535

RESUMEN

Point cloud registration is the basis of real-time environment perception for robots using 3D LiDAR and is also the key to robust simultaneous localization and mapping (SLAM) for robots. Because LiDAR point clouds are characterized by local sparseness and motion distortion, the point cloud features of coal mine roadway environments show a weak texture and degradation. Therefore, for these environments, the traditional point cloud registration method to register directly will lead to problems, such as a decline in registration accuracy, z-axis drift, and map ghosting. To solve the above problems, we propose a point cloud registration method based on IMU preintegration with the sensor characteristics of LiDAR and IMU. The system framework of this method mainly consists of four modules: IMU preintegration, point cloud preprocessing, point cloud frame matching and point cloud registration. First, IMU sensor data are introduced, and IMU linear interpolation is used to correct the motion distortion in LiDAR scanning, and the IMU preintegration error function is constructed. Second, the point cloud segmentation is performed using the ground segmentation method of RANSAC to provide additional ground constraints for the z-axis displacement and to remove the unstable flawed points from the point cloud. On this basis, the LiDAR point cloud registration error function is constructed by extracting the feature corner points and feature plane points. Finally, the Gaussian Newton solution is used to optimize the constraint relationship between the LiDAR odometry frames to minimize the error function, complete the LiDAR point cloud registration and better estimate the position and pose of the mobile robot. The experimental results show that compared with the traditional point cloud registration method, the proposed method has a higher point cloud registration accuracy, success rate and computational efficiency. The LiDAR odometry constructed using this method can better reflect the authenticity of the robot trajectory and has higher trajectory accuracy and smaller absolute position and pose error.

4.
Cereb Cortex ; 30(12): 6270-6283, 2020 11 03.
Artículo en Inglés | MEDLINE | ID: mdl-32637986

RESUMEN

Perceptual processing along the ventral visual pathway to the hippocampus (HPC) is hypothesized to be substantiated by signal transformation from retinotopic space to relational space, which represents interrelations among constituent visual elements. However, our visual perception necessarily reflects the first person's perspective based on the retinotopic space. To investigate this two-facedness of visual perception, we compared neural activities in the temporal lobe (anterior inferotemporal cortex, perirhinal and parahippocampal cortices, and HPC) between when monkeys gazed on an object and when they fixated on the screen center with an object in their peripheral vision. We found that in addition to the spatially invariant object signal, the temporal lobe areas automatically represent a large-scale background image, which specify the subject's viewing location. These results suggest that a combination of two distinct visual signals on relational space and retinotopic space may provide the first person's perspective serving for perception and presumably subsequent episodic memory.


Asunto(s)
Neuronas/fisiología , Lóbulo Temporal/fisiología , Percepción Visual/fisiología , Animales , Macaca mulatta , Masculino , Estimulación Luminosa , Vías Visuales/fisiología
5.
Sensors (Basel) ; 22(1)2021 Dec 28.
Artículo en Inglés | MEDLINE | ID: mdl-35009736

RESUMEN

Road and sidewalk detection in urban scenarios is a challenging task because of the road imperfections and high sensor data bandwidth. Traditional free space and ground filter algorithms are not sensitive enough for small height differences. Camera-based or sensor-fusion solutions are widely used to classify drivable road from sidewalk or pavement. A LIDAR sensor contains all the necessary information from which the feature extraction can be done. Therefore, this paper focuses on LIDAR-based feature extraction. For road and sidewalk detection, the current paper presents a real-time (20 Hz+) solution. This solution can also be used for local path planning. Sidewalk edge detection is the combination of three algorithms working parallelly. To validate the result, the de facto standard benchmark dataset, KITTI, was used alongside our measurements. The data and the source code to reproduce the results are shared publicly on our GitHub repository.


Asunto(s)
Algoritmos , Vehículos Autónomos
6.
Sensors (Basel) ; 20(13)2020 Jul 03.
Artículo en Inglés | MEDLINE | ID: mdl-32635370

RESUMEN

We propose a completely unsupervised approach to simultaneously estimate scene depth, ego-pose, ground segmentation and ground normal vector from only monocular RGB video sequences. In our approach, estimation for different scene structures can mutually benefit each other by the joint optimization. Specifically, we use the mutual information loss to pre-train the ground segmentation network and before adding the corresponding self-learning label obtained by a geometric method. By using the static nature of the ground and its normal vector, the scene depth and ego-motion can be efficiently learned by the self-supervised learning procedure. Extensive experimental results on both Cityscapes and KITTI benchmark demonstrate the significant improvement on the estimation accuracy for both scene depth and ego-pose by our approach. We also achieve an average error of about 3° for estimated ground normal vectors. By deploying our proposed geometric constraints, the IOUaccuracy of unsupervised ground segmentation is increased by 35% on the Cityscapes dataset.

7.
J Neurophysiol ; 121(5): 1917-1923, 2019 05 01.
Artículo en Inglés | MEDLINE | ID: mdl-30917072

RESUMEN

Discerning objects from their surrounds (i.e., figure-ground segmentation) in a way that guides adaptive behaviors is a fundamental task of the brain. Neurophysiological work has revealed a class of cells in the macaque visual cortex that may be ideally suited to support this neural computation: border ownership cells (Zhou H, Friedman HS, von der Heydt R. J Neurosci 20: 6594-6611, 2000). These orientation-tuned cells appear to respond conditionally to the borders of objects. A behavioral correlate supporting the existence of these cells in humans was demonstrated with two-dimensional luminance-defined objects (von der Heydt R, Macuda T, Qiu FT. J Opt Soc Am A Opt Image Sci Vis 22: 2222-2229, 2005). However, objects in our natural visual environments are often signaled by complex cues, such as motion and binocular disparity. Thus for border ownership systems to effectively support figure-ground segmentation and object depth ordering, they must have access to information from multiple depth cues with strict depth order selectivity. Here we measured in humans (of both sexes) border ownership-dependent tilt aftereffects after adaptation to figures defined by either motion parallax or binocular disparity. We find that both depth cues produce a tilt aftereffect that is selective for figure-ground depth order. Furthermore, we find that the effects of adaptation are transferable between cues, suggesting that these systems may combine depth cues to reduce uncertainty (Bülthoff HH, Mallot HA. J Opt Soc Am A 5: 1749-1758, 1988). These results suggest that border ownership mechanisms have strict depth order selectivity and access to multiple depth cues that are jointly encoded, providing compelling psychophysical support for their role in figure-ground segmentation in natural visual environments. NEW & NOTEWORTHY Figure-ground segmentation is a critical function that may be supported by "border ownership" neural systems that conditionally respond to object borders. We measured border ownership-dependent tilt aftereffects to figures defined by motion parallax or binocular disparity and found aftereffects for both cues. These effects were transferable between cues but selective for figure-ground depth order, suggesting that the neural systems supporting figure-ground segmentation have strict depth order selectivity and access to multiple depth cues that are jointly encoded.


Asunto(s)
Disparidad Visual , Adaptación Fisiológica , Adulto , Señales (Psicología) , Femenino , Humanos , Masculino , Percepción de Movimiento , Visión Binocular
8.
Brain Topogr ; 31(2): 202-217, 2018 03.
Artículo en Inglés | MEDLINE | ID: mdl-28808817

RESUMEN

Figure recognition process: From the coarse configuration standing from the background to the closure of a meaningful shape, was investigated by ERP technique. ERP components at different latencies from stimulus onset allowed to tap into the figure recognition process at discrete time-points when different cognitive operations take place. In this study, we present two experiments where the support-ratio (SR) of illusory figures was manipulated to vary continuously the recognition of geometrical figures. In the first experiment three shapes were used to vary the SR and the P1 component (80-130 ms) was modulated by the configuration-effect explained, in part for the first time, with the unbalanced physical stimulation between upper and lower visual field. In the second experiment, we used one shape and varied systematically the SR in a discrimination task. The N1 (130-180 ms) and the N2 (230-270 ms) were modulated by two effects: The Ic-effect, represented by the N1, and the closure-effect, represented by the N2, being larger when the SR was small and the discrimination more difficult with respect to large SRs and easier discrimination. These results showed that figure recognition proceeded from the coarse parsing of the visual scene (configuration-effect), through the completion of a set of illusory borders (Ic-effect) to the final assembling of a meaningful shape (closure-effect).


Asunto(s)
Potenciales Evocados Visuales/fisiología , Percepción de Forma/fisiología , Ilusiones/fisiología , Reconocimiento Visual de Modelos/fisiología , Adolescente , Adulto , Electroencefalografía , Femenino , Humanos , Masculino , Estimulación Luminosa/métodos , Campos Visuales/fisiología , Adulto Joven
9.
Cereb Cortex ; 26(10): 3964-76, 2016 10.
Artículo en Inglés | MEDLINE | ID: mdl-27522074

RESUMEN

Segregation of images into figures and background is fundamental for visual perception. Cortical neurons respond more strongly to figural image elements than to background elements, but the mechanisms of figure-ground modulation (FGM) are only partially understood. It is unclear whether FGM in early and mid-level visual cortex is caused by an enhanced response to the figure, a suppressed response to the background, or both.We studied neuronal activity in areas V1 and V4 in monkeys performing a texture segregation task. We compared texture-defined figures with homogeneous textures and found an early enhancement of the figure representation, and a later suppression of the background. Across neurons, the strength of figure enhancement was independent of the strength of background suppression.We also examined activity in the different V1 layers. Both figure enhancement and ground suppression were strongest in superficial and deep layers and weaker in layer 4. The current-source density profiles suggested that figure enhancement was caused by stronger synaptic inputs in feedback-recipient layers 1, 2, and 5 and ground suppression by weaker inputs in these layers, suggesting an important role for feedback connections from higher level areas. These results provide new insights into the mechanisms for figure-ground organization.


Asunto(s)
Neuronas/fisiología , Reconocimiento Visual de Modelos/fisiología , Corteza Visual/fisiología , Potenciales de Acción , Animales , Atención/fisiología , Electrodos Implantados , Medidas del Movimiento Ocular , Haplorrinos , Pruebas Neuropsicológicas , Estimulación Luminosa , Procesamiento de Señales Asistido por Computador
10.
Cereb Cortex ; 26(5): 1997-2005, 2016 May.
Artículo en Inglés | MEDLINE | ID: mdl-25681421

RESUMEN

Many theories of visual object perception assume the visual system initially extracts borders between objects and their background and then "fills in" color to the resulting object surfaces. We investigated the transformation of chromatic signals across the human ventral visual stream, with particular interest in distinguishing representations of object surface color from representations of chromatic signals reflecting the retinal input. We used fMRI to measure brain activity while participants viewed figure-ground stimuli that differed either in the position or in the color contrast polarity of the foreground object (the figure). Multivariate pattern analysis revealed that classifiers were able to decode information about which color was presented at a particular retinal location from early visual areas, whereas regions further along the ventral stream exhibited biases for representing color as part of an object's surface, irrespective of its position on the retina. Additional analyses showed that although activity in V2 contained strong chromatic contrast information to support the early parsing of objects within a visual scene, activity in this area also signaled information about object surface color. These findings are consistent with the view that mechanisms underlying scene segmentation and the binding of color to object surfaces converge in V2.


Asunto(s)
Percepción de Color/fisiología , Visión de Colores/fisiología , Percepción de Forma/fisiología , Sensibilidad de Contraste , Femenino , Humanos , Masculino , Análisis Multivariante , Estimulación Luminosa , Propiedades de Superficie
11.
Sensors (Basel) ; 15(9): 21931-56, 2015 Aug 31.
Artículo en Inglés | MEDLINE | ID: mdl-26404290

RESUMEN

The quick and accurate understanding of the ambient environment, which is composed of road curbs, vehicles, pedestrians, etc., is critical for developing intelligent vehicles. The road elements included in this work are road curbs and dynamic road obstacles that directly affect the drivable area. A framework for the online modeling of the driving environment using a multi-beam LIDAR, i.e., a Velodyne HDL-64E LIDAR, which describes the 3D environment in the form of a point cloud, is reported in this article. First, ground segmentation is performed via multi-feature extraction of the raw data grabbed by the Velodyne LIDAR to satisfy the requirement of online environment modeling. Curbs and dynamic road obstacles are detected and tracked in different manners. Curves are fitted for curb points, and points are clustered into bundles whose form and kinematics parameters are calculated. The Kalman filter is used to track dynamic obstacles, whereas the snake model is employed for curbs. Results indicate that the proposed framework is robust under various environments and satisfies the requirements for online processing.

12.
Front Plant Sci ; 15: 1393592, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38957596

RESUMEN

The nonuniform distribution of fruit tree canopies in space poses a challenge for precision management. In recent years, with the development of Structure from Motion (SFM) technology, unmanned aerial vehicle (UAV) remote sensing has been widely used to measure canopy features in orchards to balance efficiency and accuracy. A pipeline of canopy volume measurement based on UAV remote sensing was developed, in which RGB and digital surface model (DSM) orthophotos were constructed from captured RGB images, and then the canopy was segmented using U-Net, OTSU, and RANSAC methods, and the volume was calculated. The accuracy of the segmentation and the canopy volume measurement were compared. The results show that the U-Net trained with RGB and DSM achieves the best accuracy in the segmentation task, with mean intersection of concatenation (MIoU) of 84.75% and mean pixel accuracy (MPA) of 92.58%. However, in the canopy volume estimation task, the U-Net trained with DSM only achieved the best accuracy with Root mean square error (RMSE) of 0.410 m3, relative root mean square error (rRMSE) of 6.40%, and mean absolute percentage error (MAPE) of 4.74%. The deep learning-based segmentation method achieved higher accuracy in both the segmentation task and the canopy volume measurement task. For canopy volumes up to 7.50 m3, OTSU and RANSAC achieve an RMSE of 0.521 m3 and 0.580 m3, respectively. Therefore, in the case of manually labeled datasets, the use of U-Net to segment the canopy region can achieve higher accuracy of canopy volume measurement. If it is difficult to cover the cost of data labeling, ground segmentation using partitioned OTSU can yield more accurate canopy volumes than RANSAC.

13.
Atten Percept Psychophys ; 84(7): 2255-2270, 2022 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-35610411

RESUMEN

Stimulus and response features are linked together into an event file when a response is made towards a stimulus. If some or all linked features repeat, the whole event file (including the previous response) is retrieved, thereby affecting current performance (as measured in so-called binding effects). Applying the figure-ground segmentation principle to such action control experiments, previous research showed that only stimulus features that have a figure-like character led to binding effects, while features in the background did not. Against the background of recent theorizing, integration and retrieval are discussed as separate processes that independently contribute to binding effects (BRAC framework). Thus, previous research did not specify whether figure-ground manipulations exert their modulating influence on integration and/or retrieval. We tested this in three experiments. Participants worked through a sequential distractor-response binding (DRB) task, allowing measurement of binding effects between responses and distractor (color) features. Importantly, we manipulated whether the distractor color was presented as a background feature or as a figure feature. In contrast to previous experiments, we applied this manipulation only to prime displays (Experiment 1), only to probe display (Experiment 2), or varied the figure-ground manipulation orthogonally for primes and probes (Experiment 3). Together the results of all three experiments suggest that figure-ground segmentation affects DRB effects on top of encoding specificity, and that especially the retrieval process is affected by this manipulation.


Asunto(s)
Atención , Percepción , Atención/fisiología , Humanos , Tiempo de Reacción/fisiología
14.
Front Behav Neurosci ; 15: 756801, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-34938164

RESUMEN

Recent work has shown that the medial temporal lobe (MTL), including the hippocampus (HPC) and its surrounding limbic cortices, plays a role in scene perception in addition to episodic memory. The two basic factors of scene perception are the object ("what") and location ("where"). In this review, we first summarize the anatomical knowledge related to visual inputs to the MTL and physiological studies examining object-related information processed along the ventral pathway briefly. Thereafter, we discuss the space-related information, the processing of which was unclear, presumably because of its multiple aspects and a lack of appropriate task paradigm in contrast to object-related information. Based on recent electrophysiological studies using non-human primates and the existing literature, we proposed the "reunification theory," which explains brain mechanisms which construct object-location signals at each gaze. In this reunification theory, the ventral pathway signals a large-scale background image of the retina at each gaze position. This view-center background signal reflects the first person's perspective and specifies the allocentric location in the environment by similarity matching between images. The spatially invariant object signal and view-center background signal, both of which are derived from the same retinal image, are integrated again (i.e., reunification) along the ventral pathway-MTL stream, particularly in the perirhinal cortex. The conjunctive signal, which represents a particular object at a particular location, may play a role in scene perception in the HPC as a key constituent element of an entire scene.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA