Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 10 de 10
Filtrar
1.
Opt Express ; 31(2): 2538-2551, 2023 Jan 16.
Artículo en Inglés | MEDLINE | ID: mdl-36785265

RESUMEN

One of the open challenges in lensless imaging is understanding how well they resolve scenes in three dimensions. The measurement model underlying prior lensless imagers lacks special structures that facilitate deeper analysis; thus, a theoretical study of the achievable spatio-axial resolution has been lacking. This paper provides such a theoretical framework by analyzing a generalization of a mask-based lensless camera, where the sensor captures z-stacked measurements acquired by moving the sensor relative to an attenuating mask. We show that the z-stacked measurements are related to the scene's volumetric albedo function via a three-dimensional convolutional operator. The specifics of this convolution, and its Fourier transform, allow us to fully characterize the spatial and axial resolving power of the camera, including its dependence on the mask. Since z-stacked measurements are a superset of those made by previously-studied lensless systems, these results provide an upper bound for their performance. We numerically evaluate the theory and its implications using simulations.

2.
Sensors (Basel) ; 22(24)2022 Dec 17.
Artículo en Inglés | MEDLINE | ID: mdl-36560332

RESUMEN

In this paper, we present a framework to learn illumination patterns to improve the quality of signal recovery for coded diffraction imaging. We use an alternating minimization-based phase retrieval method with a fixed number of iterations as the iterative method. We represent the iterative phase retrieval method as an unrolled network with a fixed number of layers where each layer of the network corresponds to a single step of iteration, and we minimize the recovery error by optimizing over the illumination patterns. Since the number of iterations/layers is fixed, the recovery has a fixed computational cost. Extensive experimental results on a variety of datasets demonstrate that our proposed method significantly improves the quality of image reconstruction at a fixed computational cost with illumination patterns learned only using a small number of training images.


Asunto(s)
Algoritmos , Diagnóstico por Imagen , Procesamiento de Imagen Asistido por Computador/métodos
3.
Sensors (Basel) ; 22(14)2022 Jul 18.
Artículo en Inglés | MEDLINE | ID: mdl-35891033

RESUMEN

In current decades, significant advancements in robotics engineering and autonomous vehicles have improved the requirement for precise depth measurements. Depth estimation (DE) is a traditional task in computer vision that can be appropriately predicted by applying numerous procedures. This task is vital in disparate applications such as augmented reality and target tracking. Conventional monocular DE (MDE) procedures are based on depth cues for depth prediction. Various deep learning techniques have demonstrated their potential applications in managing and supporting the traditional ill-posed problem. The principal purpose of this paper is to represent a state-of-the-art review of the current developments in MDE based on deep learning techniques. For this goal, this paper tries to highlight the critical points of the state-of-the-art works on MDE from disparate aspects. These aspects include input data shapes and training manners such as supervised, semi-supervised, and unsupervised learning approaches in combination with applying different datasets and evaluation indicators. At last, limitations regarding the accuracy of the DL-based MDE models, computational time requirements, real-time inference, transferability, input images shape and domain adaptation, and generalization are discussed to open new directions for future research.


Asunto(s)
Realidad Aumentada , Aprendizaje Profundo , Predicción
4.
Artículo en Inglés | MEDLINE | ID: mdl-39388325

RESUMEN

Multimodal learning seeks to utilize data from multiple sources to improve the overall performance of downstream tasks. It is desirable for redundancies in the data to make multimodal systems robust to missing or corrupted observations in some correlated modalities. However, we observe that the performance of several existing multimodal networks significantly deteriorates if one or multiple modalities are absent at test time. To enable robustness to missing modalities, we propose a simple and parameter-efficient adaptation procedure for pretrained multimodal networks. In particular, we exploit modulation of intermediate features to compensate for the missing modalities. We demonstrate that such adaptation can partially bridge performance drop due to missing modalities and outperform independent, dedicated networks trained for the available modality combinations in some cases. The proposed adaptation requires extremely small number of parameters (e.g., fewer than 1% of the total parameters) and applicable to a wide range of modality combinations and tasks. We conduct a series of experiments to highlight the missing modality robustness of our proposed method on five different multimodal tasks across seven datasets. Our proposed method demonstrates versatility across various tasks and datasets, and outperforms existing methods for robust multimodal learning with missing modalities.

5.
IEEE Trans Pattern Anal Mach Intell ; 46(7): 4684-4701, 2024 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-38285590

RESUMEN

Conventional cameras capture image irradiance (RAW) on a sensor and convert it to RGB images using an image signal processor (ISP). The images can then be used for photography or visual computing tasks in a variety of applications, such as public safety surveillance and autonomous driving. One can argue that since RAW images contain all the captured information, the conversion of RAW to RGB using an ISP is not necessary for visual computing. In this paper, we propose a novel ρ-Vision framework to perform high-level semantic understanding and low-level compression using RAW images without the ISP subsystem used for decades. Considering the scarcity of available RAW image datasets, we first develop an unpaired CycleR2R network based on unsupervised CycleGAN to train modular unrolled ISP and inverse ISP (invISP) models using unpaired RAW and RGB images. We can then flexibly generate simulated RAW images (simRAW) using any existing RGB image dataset and finetune different models originally trained in the RGB domain to process real-world camera RAW images. We demonstrate object detection and image compression capabilities in RAW-domain using RAW-domain YOLOv3 and RAW image compressor (RIC) on camera snapshots. Quantitative results reveal that RAW-domain task inference provides better detection accuracy and compression efficiency compared to that in the RGB domain. Furthermore, the proposed ρ-Vision generalizes across various camera sensors and different task-specific models. An added benefit of employing the ρ-Vision is the elimination of the need for ISP, leading to potential reductions in computations and processing times.

6.
Magn Reson Med ; 70(3): 800-12, 2013 Sep.
Artículo en Inglés | MEDLINE | ID: mdl-23132400

RESUMEN

Accelerated magnetic resonance imaging techniques reduce signal acquisition time by undersampling k-space. A fundamental problem in accelerated magnetic resonance imaging is the recovery of quality images from undersampled k-space data. Current state-of-the-art recovery algorithms exploit the spatial and temporal structures in underlying images to improve the reconstruction quality. In recent years, compressed sensing theory has helped formulate mathematical principles and conditions that ensure recovery of (structured) sparse signals from undersampled, incoherent measurements. In this article, a new recovery algorithm, motion-adaptive spatio-temporal regularization, is presented that uses spatial and temporal structured sparsity of MR images in the compressed sensing framework to recover dynamic MR images from highly undersampled k-space data. In contrast to existing algorithms, our proposed algorithm models temporal sparsity using motion-adaptive linear transformations between neighboring images. The efficiency of motion-adaptive spatio-temporal regularization is demonstrated with experiments on cardiac magnetic resonance imaging for a range of reduction factors. Results are also compared with k-t FOCUSS with motion estimation and compensation-another recently proposed recovery algorithm for dynamic magnetic resonance imaging. .


Asunto(s)
Imagen por Resonancia Magnética/métodos , Algoritmos , Modelos Teóricos , Movimiento (Física) , Análisis Espacio-Temporal
7.
ACS Appl Mater Interfaces ; 15(14): 18244-18251, 2023 Apr 12.
Artículo en Inglés | MEDLINE | ID: mdl-37010900

RESUMEN

The rapid characterization of nanoparticles for morphological information such as size and shape is essential for material synthesis as they are the determining factors for the optical, mechanical, and chemical properties and related applications. In this paper, we report a computational imaging platform to characterize nanoparticle size and morphology under conventional optical microscopy. We established a machine learning model based on a series of images acquired by through-focus scanning optical microscopy (TSOM) on a conventional optical microscope. This model predicts the size of silver nanocubes with an estimation error below 5% on individual particles. At the ensemble level, the estimation error is 1.6% for the averaged size and 0.4 nm for the standard deviation. The method can also identify the tip morphology of silver nanowires from the mix of sharp-tip and blunt-tip samples at an accuracy of 82%. Furthermore, we demonstrated online monitoring for the evolution of the size distribution of nanoparticles during synthesis. This method can be potentially extended to more complicated nanomaterials such as anisotropic and dielectric nanoparticles.

8.
Artículo en Inglés | MEDLINE | ID: mdl-35649169

RESUMEN

Controlling the nanoscale light-matter interaction using superfocusing hybrid photonic-plasmonic devices has attracted significant research interest in tackling existing challenges, including converting efficiencies, working bandwidths, and manufacturing complexities. With the growth in demand for efficient photonic-plasmonic input-output interfaces to improve plasmonic device performances, sophisticated designs with multiple optimization parameters are required, which comes with an unaffordable computation cost. Machine learning methods can significantly reduce the cost of computations compared to numerical simulations, but the input-output dimension mismatch remains a challenging problem. Here, we introduce a physics-guided two-stage machine learning network that uses the improved coupled-mode theory for optical waveguides to guide the learning module and improve the accuracy of predictive engines to 98.5%. A near-unity coupling efficiency with symmetry-breaking selectivity is predicted by the inverse design. By fabricating photonic-plasmonic couplers using the predicted profiles, we demonstrate that the excitation efficiency of 83% on the radially polarized surface plasmon mode can be achieved, which paves the way for super-resolution optical imaging.

9.
IEEE Trans Pattern Anal Mach Intell ; 43(10): 3275-3291, 2021 Oct.
Artículo en Inglés | MEDLINE | ID: mdl-32248090

RESUMEN

This paper presents a dual camera system for high spatiotemporal resolution (HSTR) video acquisition, where one camera shoots a video with high spatial resolution and low frame rate (HSR-LFR) and another one captures a low spatial resolution and high frame rate (LSR-HFR) video. Our main goal is to combine videos from LSR-HFR and HSR-LFR cameras to create an HSTR video. We propose an end-to-end learning framework, AWnet, mainly consisting of a FlowNet and a FusionNet that learn an adaptive weighting function in pixel domain to combine inputs in a frame recurrent fashion. To improve the reconstruction quality for cameras used in reality, we also introduce noise regularization under the same framework. Our method has demonstrated noticeable performance gains in terms of both objective PSNR measurement in simulation with different publicly available video and light-field datasets and subjective evaluation with real data captured by dual iPhone 7 and Grasshopper3 cameras. Ablation studies are further conducted to investigate and explore various aspects, such as reference structure, camera parallax, exposure time, etc) of our system to fully understand its capability for potential applications.

10.
IEEE Trans Pattern Anal Mach Intell ; 42(7): 1606-1617, 2020 07.
Artículo en Inglés | MEDLINE | ID: mdl-32305898

RESUMEN

Lensless cameras, while extremely useful for imaging in constrained scenarios, struggle with resolving scenes with large depth variations. To resolve this, we propose imaging with a set of mask patterns displayed on a programmable mask, and introduce a computational focusing operator that helps to resolve the depth of scene points. As a result, the proposed imager can resolve dense scenes with large depth variations, allowing for more practical applications of lensless cameras. We also present a fast reconstruction algorithm for scene at multiple depths that reduces reconstruction time by two orders of magnitude. Finally, we build a prototype to show the proposed method improves both image quality and depth resolution of lensless cameras.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA