Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 29
Filtrar
1.
Sensors (Basel) ; 24(2)2024 Jan 19.
Artículo en Inglés | MEDLINE | ID: mdl-38276341

RESUMEN

Industrial-quality inspections, particularly those leveraging AI, require significant amounts of training data. In fields like injection molding, producing a multitude of defective parts for such data poses environmental and financial challenges. Synthetic training data emerge as a potential solution to address these concerns. Although the creation of realistic synthetic 2D images from 3D models of injection-molded parts involves numerous rendering parameters, the current literature on the generation and application of synthetic data in industrial-quality inspection scarcely addresses the impact of these parameters on AI efficacy. In this study, we delve into some of these key parameters, such as camera position, lighting, and computational noise, to gauge their effect on AI performance. By utilizing Blender software, we procedurally introduced the "flash" defect on a 3D model sourced from a CAD file of an injection-molded part. Subsequently, with Blender's Cycles rendering engine, we produced datasets for each parameter variation. These datasets were then used to train a pre-trained EfficientNet-V2 for the binary classification of the "flash" defect. Our results indicate that while noise is less critical, using a range of noise levels in training can benefit model adaptability and efficiency. Variability in camera positioning and lighting conditions was found to be more significant, enhancing model performance even when real-world conditions mirror the controlled synthetic environment. These findings suggest that incorporating diverse lighting and camera dynamics is beneficial for AI applications, regardless of the consistency in real-world operational settings.

2.
Sensors (Basel) ; 24(7)2024 Apr 03.
Artículo en Inglés | MEDLINE | ID: mdl-38610501

RESUMEN

Multimodal sensors capture and integrate diverse characteristics of a scene to maximize information gain. In optics, this may involve capturing intensity in specific spectra or polarization states to determine factors such as material properties or an individual's health conditions. Combining multimodal camera data with shape data from 3D sensors is a challenging issue. Multimodal cameras, e.g., hyperspectral cameras, or cameras outside the visible light spectrum, e.g., thermal cameras, lack strongly in terms of resolution and image quality compared with state-of-the-art photo cameras. In this article, a new method is demonstrated to superimpose multimodal image data onto a 3D model created by multi-view photogrammetry. While a high-resolution photo camera captures a set of images from varying view angles to reconstruct a detailed 3D model of the scene, low-resolution multimodal camera(s) simultaneously record the scene. All cameras are pre-calibrated and rigidly mounted on a rig, i.e., their imaging properties and relative positions are known. The method was realized in a laboratory setup consisting of a professional photo camera, a thermal camera, and a 12-channel multispectral camera. In our experiments, an accuracy better than one pixel was achieved for the data fusion using multimodal superimposition. Finally, application examples of multimodal 3D digitization are demonstrated, and further steps to system realization are discussed.

3.
Sensors (Basel) ; 23(16)2023 Aug 08.
Artículo en Inglés | MEDLINE | ID: mdl-37631565

RESUMEN

The projection of a point cloud onto a 2D camera image is relevant in the case of various image analysis and enhancement tasks, e.g., (i) in multimodal image processing for data fusion, (ii) in robotic applications and in scene analysis, and (iii) for deep neural networks to generate real datasets with ground truth. The challenges of the current single-shot projection methods, such as simple state-of-the-art projection, conventional, polygon, and deep learning-based upsampling methods or closed source SDK functions of low-cost depth cameras, have been identified. We developed a new way to project point clouds onto a dense, accurate 2D raster image, called Triangle-Mesh-Rasterization-Projection (TMRP). The only gaps that the 2D image still contains with our method are valid gaps that result from the physical limits of the capturing cameras. Dense accuracy is achieved by simultaneously using the 2D neighborhood information (rx,ry) of the 3D coordinates in addition to the points P(X,Y,V). In this way, a fast triangulation interpolation can be performed. The interpolation weights are determined using sub-triangles. Compared to single-shot methods, our algorithm is able to solve the following challenges. This means that: (1) no false gaps or false neighborhoods are generated, (2) the density is XYZ independent, and (3) ambiguities are eliminated. Our TMRP method is also open source, freely available on GitHub, and can be applied to almost any sensor or modality. We also demonstrate the usefulness of our method with four use cases by using the KITTI-2012 dataset or sensors with different modalities. Our goal is to improve recognition tasks and processing optimization in the perception of transparent objects for robotic manufacturing processes.

4.
Sensors (Basel) ; 23(18)2023 Sep 11.
Artículo en Inglés | MEDLINE | ID: mdl-37765862

RESUMEN

In the context of collaborative robotics, handing over hand-held objects to a robot is a safety-critical task. Therefore, a robust distinction between human hands and presented objects in image data is essential to avoid contact with robotic grippers. To be able to develop machine learning methods for solving this problem, we created the OHO (Object Hand-Over) dataset of tools and other everyday objects being held by human hands. Our dataset consists of color, depth, and thermal images with the addition of pose and shape information about the objects in a real-world scenario. Although the focus of this paper is on instance segmentation, our dataset also enables training for different tasks such as 3D pose estimation or shape estimation of objects. For the instance segmentation task, we present a pipeline for automated label generation in point clouds, as well as image data. Through baseline experiments, we show that these labels are suitable for training an instance segmentation to distinguish hands from objects on a per-pixel basis. Moreover, we present qualitative results for applying our trained model in a real-world application.


Asunto(s)
Robótica , Humanos , Aprendizaje Automático , Extremidad Superior
5.
Sensors (Basel) ; 23(20)2023 Oct 18.
Artículo en Inglés | MEDLINE | ID: mdl-37896662

RESUMEN

Estimating depth from images is a common technique in 3D perception. However, dealing with non-Lambertian materials, e.g., transparent or specular, is still nowadays an open challenge. However, to overcome this challenge with deep stereo matching networks or monocular depth estimation, data sets with non-Lambertian objects are mandatory. Currently, only few real-world data sets are available. This is due to the high effort and time-consuming process of generating these data sets with ground truth. Currently, transparent objects must be prepared, e.g., painted or powdered, or an opaque twin of the non-Lambertian object is needed. This makes data acquisition very time consuming and elaborate. We present a new measurement principle for how to generate a real data set of transparent and specular surfaces without object preparation techniques, which greatly reduces the effort and time required for data collection. For this purpose, we use a thermal 3D sensor as a reference system, which allows the 3D detection of transparent and reflective surfaces without object preparation. In addition, we publish the first-ever real stereo data set, called TranSpec3D, where ground truth disparities without object preparation were generated using this measurement principle. The data set contains 110 objects and consists of 148 scenes, each taken in different lighting environments, which increases the size of the data set and creates different reflections on the surface. We also show the advantages and disadvantages of our measurement principle and data set compared to the Booster data set (generated with object preparation), as well as the current limitations of our novel method.

6.
Opt Express ; 30(22): 39534-39543, 2022 Oct 24.
Artículo en Inglés | MEDLINE | ID: mdl-36298903

RESUMEN

For three-dimensional (3D) measurement of object surface and shape by pattern projection systems, we used a hybrid projection system, i.e., a combination of a projection lens and a transmissive freeform to generate an aperiodic sinusoidal fringe pattern. Such a freeform effects a light redistribution, thus leading to an effective and low-loss pattern projection, as it increases the total transmission intensity of the system and has less power dissipation than classical projection systems. In this paper, we present the conception and realization of the measurement setup of a transmissive fringe projection system. We compare the characteristics of the generated intensity distribution with the classical system based on GOBO (GOes Before Optics) projection and show measurement results of different surface shapes, recorded with the new system.

7.
Opt Express ; 30(13): 22590-22607, 2022 Jun 20.
Artículo en Inglés | MEDLINE | ID: mdl-36224953

RESUMEN

Close-range 3D sensors based on the structured light principle have a constrained measuring range due to their depth of field (DOF). Focus stacking is a method to extend the DOF. The additional time to change the focus is a drawback in high-speed measurements. In our research, the method of chromatic focus stacking was applied to a high-speed 3D sensor with 180 fps frame rate. The extended DOF was evaluated by the distance-dependent 3D resolution derived from the 3D-MTF of a tilted edge. The conventional DOF of 14 mm was extended to 21 mm by stacking two foci at 455 and 520 nm wavelength. The 3D sensor allowed shape measurements with extended DOF within 44 ms.

8.
Sensors (Basel) ; 22(19)2022 Oct 05.
Artículo en Inglés | MEDLINE | ID: mdl-36236639

RESUMEN

Geometrical camera modeling is the precondition for 3D-reconstruction tasks using photogrammetric sensor systems. The purpose of this study is to describe an approach for possible accuracy improvements by using the ray-based-camera model. The relations between the common pinhole and the generally valid ray-based-camera model are shown. A new approach to the implementation and calibration of the ray-based-camera model is introduced. Using a simple laboratory setup consisting of two cameras and a projector, experimental measurements were performed. The experiments and results showed the possibility of easily transforming the common pinhole model into a ray-based model and of performing calibration using the ray-based model. These initial results show the model's potential for considerable accuracy improvements, especially for sensor systems using wide-angle lenses or with deep 3D measurements. This study presents several approaches for further improvements to and the practical usage of high-precision optical 3D measurements.

9.
Appl Opt ; 60(8): 2362-2371, 2021 Mar 10.
Artículo en Inglés | MEDLINE | ID: mdl-33690336

RESUMEN

Three-dimensional (3D) shape measurement systems based on diffuse reflection of projected structured light do not deliver reliable data when measuring glossy, transparent, absorbent, or translucent objects. In recent years, we have developed a method based on stereo recording with infrared cameras and projection of areal aperiodic sinusoidal thermal patterns to detect such objects. However, the measurements took longer than 10 s, up to minutes; moreover, the measurement accuracy was improvable. Now, we have succeeded in both drastically reducing measurement time and significantly increasing measurement quality. This finally provides a technique for reliably measuring transparent objects, e.g., in series production. We demonstrate measurement examples achieved within 1 s and with 3D standard deviations less than 10 µm.

10.
Sensors (Basel) ; 21(16)2021 Aug 23.
Artículo en Inglés | MEDLINE | ID: mdl-34451117

RESUMEN

This paper presents an application of neural networks operating on multimodal 3D data (3D point cloud, RGB, thermal) to effectively and precisely segment human hands and objects held in hand to realize a safe human-robot object handover. We discuss the problems encountered in building a multimodal sensor system, while the focus is on the calibration and alignment of a set of cameras including RGB, thermal, and NIR cameras. We propose the use of a copper-plastic chessboard calibration target with an internal active light source (near-infrared and visible light). By brief heating, the calibration target could be simultaneously and legibly captured by all cameras. Based on the multimodal dataset captured by our sensor system, PointNet, PointNet++, and RandLA-Net are utilized to verify the effectiveness of applying multimodal point cloud data for hand-object segmentation. These networks were trained on various data modes (XYZ, XYZ-T, XYZ-RGB, and XYZ-RGB-T). The experimental results show a significant improvement in the segmentation performance of XYZ-RGB-T (mean Intersection over Union: 82.8% by RandLA-Net) compared with the other three modes (77.3% by XYZ-RGB, 35.7% by XYZ-T, 35.7% by XYZ), in which it is worth mentioning that the Intersection over Union for the single class of hand achieves 92.6%.


Asunto(s)
Procedimientos Quirúrgicos Robotizados , Robótica , Algoritmos , Humanos , Imagenología Tridimensional , Imagen Multimodal
11.
Opt Express ; 26(18): 23366-23379, 2018 Sep 03.
Artículo en Inglés | MEDLINE | ID: mdl-30184988

RESUMEN

Measuring the shape (coordinates x, y, z ) and spectral characteristics (wavelength-dependent reflectance R (λi)) of macroscopic objects as a function of time (t) is of great interest in areas such as medical imaging, precision agriculture, or optical sorting. Here, we present an approach that allows to determine all these quantities with high resolution and accuracy, enabling measurement in five dimensions. We call this approach 5D hyperspectral imaging. We describe the design and implementation of a 5D sensor operating in the visible to near-infrared spectral range, which provides excellent spatial and spectral resolution, great depth accuracy, and high frame rates. The results of various experiments strongly indicate the great benefit of the new technology.

12.
Appl Opt ; 56(8): 2162-2170, 2017 Mar 10.
Artículo en Inglés | MEDLINE | ID: mdl-28375301

RESUMEN

In order to perform high-speed three-dimensional (3D) shape measurements with structured light systems, high-speed projectors are required. One possibility is an array projector, which allows pattern projection at several tens of kilohertz by switching on and off the LEDs of various slide projectors. The different projection centers require a separate analysis, as the intensity received by the cameras depends on the projection direction and the object's bidirectional reflectance distribution function (BRDF). In this contribution, we investigate the BRDF-dependent errors of array-projection-based 3D sensors and propose an error compensation process.

13.
Appl Opt ; 54(35): 10541-51, 2015 Dec 10.
Artículo en Inglés | MEDLINE | ID: mdl-26836883

RESUMEN

The demand for optically reconstructing the three-dimensional (3D) surface shape of moving objects or deformation processes makes the development of high-speed projectors necessary. Our 3D sensor containing an array projector can achieve frame rates of several tens of kilohertz and is based on the projection of aperiodic sinusoidal fringes. This approach is compared with phase-shifting fringe projection as probably the most widely used technique. Theoretical considerations as well as extensive simulations are conducted to derive criteria for the design of optimal sequences of aperiodic sinusoidal fringes and to compare the number of patterns of both approaches necessary for comparable accuracies.


Asunto(s)
Imagenología Tridimensional/métodos , Simulación por Computador , Imagenología Tridimensional/estadística & datos numéricos , Modelos Teóricos , Fenómenos Ópticos , Reconocimiento de Normas Patrones Automatizadas/métodos , Reconocimiento de Normas Patrones Automatizadas/estadística & datos numéricos , Factores de Tiempo
14.
Sensors (Basel) ; 16(1)2015 Dec 23.
Artículo en Inglés | MEDLINE | ID: mdl-26703624

RESUMEN

In this work we show the principle of optical 3D surface measurements based on the fringe projection technique for underwater applications. The challenges of underwater use of this technique are shown and discussed in comparison with the classical application. We describe an extended camera model which takes refraction effects into account as well as a proposal of an effective, low-effort calibration procedure for underwater optical stereo scanners. This calibration technique combines a classical air calibration based on the pinhole model with ray-based modeling and requires only a few underwater recordings of an object of known length and a planar surface. We demonstrate a new underwater 3D scanning device based on the fringe projection technique. It has a weight of about 10 kg and the maximal water depth for application of the scanner is 40 m. It covers an underwater measurement volume of 250 mm × 200 mm × 120 mm. The surface of the measurement objects is captured with a lateral resolution of 150 µm in a third of a second. Calibration evaluation results are presented and examples of first underwater measurements are given.


Asunto(s)
Monitoreo del Ambiente/métodos , Imagenología Tridimensional/métodos , Algoritmos , Monitoreo del Ambiente/instrumentación , Hidrobiología , Imagenología Tridimensional/instrumentación , Agua/química
15.
Opt Express ; 22(11): 12982-93, 2014 Jun 02.
Artículo en Inglés | MEDLINE | ID: mdl-24921495

RESUMEN

We present Terahertz (THz) imaging with a 1D multichannel time-domain spectroscopy (TDS) system which operates with a photoconductive array of 15 detection channels excited by a 1030 nm femtosecond fiber laser. The emitter and detector are photoconductive antennas based on InGaAs/InAlAs multi-layer heterostructures (MLHS). We characterized the THz optics and the resolution of the system. The performance is demonstrated by the multichannel imaging of two samples. A simultaneous measurement of 15 THz pulses with a pixel pitch of 1 mm increases the measurement speed of the TDS system by factor 15.

16.
J Imaging ; 10(3)2024 Mar 21.
Artículo en Inglés | MEDLINE | ID: mdl-38535153

RESUMEN

Since 3D sensors became popular, imaged depth data are easier to obtain in the consumer sector. In applications such as defect localization on industrial objects or mass/volume estimation, precise depth data is important and, thus, benefits from the usage of multiple information sources. However, a combination of RGB images and depth images can not only improve our understanding of objects, capacitating one to gain more information about objects but also enhance data quality. Combining different camera systems using data fusion can enable higher quality data since disadvantages can be compensated. Data fusion itself consists of data preparation and data registration. A challenge in data fusion is the different resolutions of sensors. Therefore, up- and downsampling algorithms are needed. This paper compares multiple up- and downsampling methods, such as different direct interpolation methods, joint bilateral upsampling (JBU), and Markov random fields (MRFs), in terms of their potential to create RGB-D images and improve the quality of depth information. In contrast to the literature in which imaging systems are adjusted to acquire the data of the same section simultaneously, the laboratory setup in this study was based on conveyor-based optical sorting processes, and therefore, the data were acquired at different time periods and different spatial locations. Data assignment and data cropping were necessary. In order to evaluate the results, root mean square error (RMSE), signal-to-noise ratio (SNR), correlation (CORR), universal quality index (UQI), and the contour offset are monitored. With JBU outperforming the other upsampling methods, achieving a meanRMSE = 25.22, mean SNR = 32.80, mean CORR = 0.99, and mean UQI = 0.97.

17.
J Biomed Opt ; 29(Suppl 3): S33309, 2024 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-39170819

RESUMEN

Significance: Monitoring oxygen saturation ( SpO 2 ) is important in healthcare, especially for diagnosing and managing pulmonary diseases. Non-contact approaches broaden the potential applications of SpO 2 measurement by better hygiene, comfort, and capability for long-term monitoring. However, existing studies often encounter challenges such as lower signal-to-noise ratios and stringent environmental conditions. Aim: We aim to develop and validate a contactless SpO 2 measurement approach using 3D convolutional neural networks (3D CNN) and 3D visible-near-infrared (VIS-NIR) multimodal imaging, to offer a convenient, accurate, and robust alternative for SpO 2 monitoring. Approach: We propose an approach that utilizes a 3D VIS-NIR multimodal camera system to capture facial videos, in which SpO 2 is estimated through 3D CNN by simultaneously extracting spatial and temporal features. Our approach includes registration of multimodal images, tracking of the 3D region of interest, spatial and temporal preprocessing, and 3D CNN-based feature extraction and SpO 2 regression. Results: In a breath-holding experiment involving 23 healthy participants, we obtained multimodal video data with reference SpO 2 values ranging from 80% to 99% measured by pulse oximeter on the fingertip. The approach achieved a mean absolute error (MAE) of 2.31% and a Pearson correlation coefficient of 0.64 in the experiment, demonstrating good agreement with traditional pulse oximetry. The discrepancy of estimated SpO 2 values was within 3% of the reference SpO 2 for ∼ 80 % of all 1-s time points. Besides, in clinical trials involving patients with sleep apnea syndrome, our approach demonstrated robust performance, with an MAE of less than 2% in SpO 2 estimations compared to gold-standard polysomnography. Conclusions: The proposed approach offers a promising alternative for non-contact oxygen saturation measurement with good sensitivity to desaturation, showing potential for applications in clinical settings.


Asunto(s)
Imagenología Tridimensional , Imagen Multimodal , Redes Neurales de la Computación , Oximetría , Humanos , Oximetría/métodos , Imagen Multimodal/métodos , Adulto , Masculino , Imagenología Tridimensional/métodos , Femenino , Saturación de Oxígeno/fisiología , Adulto Joven , Espectroscopía Infrarroja Corta/métodos , Cara/diagnóstico por imagen , Cara/irrigación sanguínea , Oxígeno/sangre
18.
Sci Rep ; 14(1): 19036, 2024 08 16.
Artículo en Inglés | MEDLINE | ID: mdl-39152181

RESUMEN

With rising melanoma incidence and mortality, early detection and surgical removal of primary lesions is essential. Multispectral imaging is a new, non-invasive technique that can facilitate skin cancer detection by measuring the reflectance spectra of biological tissues. Currently, incident illumination allows little light to be reflected from deeper skin layers due to high surface reflectance. A pilot study was conducted at the University Hospital Basel to evaluate, whether multispectral imaging with direct light coupling could extract more information from deeper skin layers for more accurate dignity classification of melanocytic lesions. 27 suspicious pigmented lesions from 23 patients were included (6 melanomas, 6 dysplastic nevi, 12 melanocytic nevi, 3 other). Lesions were imaged before excision using a prototype snapshot mosaic multispectral camera with incident and direct illumination with subsequent dignity classification by a pre-trained multispectral image analysis model. Using incident light, a sensitivity of 83.3% and a specificity of 58.8% were achieved compared to dignity as determined by histopathological examination. Direct light coupling resulted in a superior sensitivity of 100% and specificity of 82.4%. Convolutional neural network classification of corresponding red, green, and blue lesion images resulted in 16.7% lower sensitivity (83.3%, 5/6 malignant lesions detected) and 20.9% lower specificity (61.5%) compared to direct light coupling with multispectral image classification. Our results show that incorporating direct light multispectral imaging into the melanoma detection process could potentially increase the accuracy of dignity classification. This newly evaluated illumination method could improve multispectral applications in skin cancer detection. Further larger studies are needed to validate the camera prototype.


Asunto(s)
Melanoma , Nevo Pigmentado , Neoplasias Cutáneas , Humanos , Melanoma/diagnóstico por imagen , Melanoma/clasificación , Melanoma/patología , Melanoma/diagnóstico , Neoplasias Cutáneas/diagnóstico por imagen , Neoplasias Cutáneas/patología , Neoplasias Cutáneas/clasificación , Neoplasias Cutáneas/diagnóstico , Femenino , Nevo Pigmentado/diagnóstico por imagen , Nevo Pigmentado/diagnóstico , Nevo Pigmentado/clasificación , Nevo Pigmentado/patología , Masculino , Persona de Mediana Edad , Adulto , Proyectos Piloto , Anciano , Melanocitos/patología , Iluminación/métodos , Procesamiento de Imagen Asistido por Computador/métodos , Sensibilidad y Especificidad
19.
Appl Opt ; 52(14): 3134-46, 2013 May 10.
Artículo en Inglés | MEDLINE | ID: mdl-23669825

RESUMEN

Phase-shifting fringe projection is an effective method to perform 3D shape measurements. Conventionally, fringe projection systems utilize a digital projector that images fringes into the measurement plane. The performance of such systems is limited to the visible spectral range, as most projectors experience technical limitations in UV or IR spectral ranges. However, for certain applications these spectral ranges are of special interest. We present a wideband fringe projector that has been developed on the basis of a picture generating beamshaping mirror. This mirror generates a sinusoidal fringe pattern in the measurement plane without any additional optical elements. Phase shifting is realized without any mechanical movement by a multichip LED. As the system is based on a single mirror, it is wavelength-independent in a wide spectral range and therefore applicable in UV and IR spectral ranges. We present the design and a realized setup of this fringe projection system and the characterization of the generated intensity distribution. Experimental results of 3D shape measurements are presented.

20.
Front Robot AI ; 10: 1120357, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-37008984

RESUMEN

The concept of Industry 4.0 brings the change of industry manufacturing patterns that become more efficient and more flexible. In response to this tendency, an efficient robot teaching approach without complex programming has become a popular research direction. Therefore, we propose an interactive finger-touch based robot teaching schema using a multimodal 3D image (color (RGB), thermal (T) and point cloud (3D)) processing. Here, the resulting heat trace touching the object surface will be analyzed on multimodal data, in order to precisely identify the true hand/object contact points. These identified contact points are used to calculate the robot path directly. To optimize the identification of the contact points we propose a calculation scheme using a number of anchor points which are first predicted by hand/object point cloud segmentation. Subsequently a probability density function is defined to calculate the prior probability distribution of true finger trace. The temperature in the neighborhood of each anchor point is then dynamically analyzed to calculate the likelihood. Experiments show that the trajectories estimated by our multimodal method have significantly better accuracy and smoothness than only by analyzing point cloud and static temperature distribution.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA