In a paper published in the journal Optics & Laser Technology, researchers introduced the image-loop neural network (ILNet), a self-supervised deep learning technique designed for single-pixel imaging (SPI) integrated with a part-based model. This innovative approach offers a versatile and efficient solution for SPI, enabling the reconstruction of high-quality images at low sampling rates. Promising outcomes have been demonstrated in both free-space and underwater scenarios.
Background
SPI is a potent technique that reconstructs two-dimensional (2D) images from one-dimensional (1D) intensity signals using the correlation of classical or quantum light. Its applications span various domains, including microscopic, X-ray, and lidar imaging, primarily due to its high sensitivity in challenging conditions with low light, long distances, and high scattering. However, the requirement for substantial data for high-quality image reconstruction limits SPI's real-time imaging capabilities.
To address this constraint, researchers have turned to deep learning (DL) approaches, leveraging neural networks' capacity to extract underlying rules and representations from data. DL-based SPI methods, such as underwater ghost imaging-based generative adversarial networks (UGI-GAN), multi-scale GAN (MsGAN), and ghost imaging using deep learning (GIDL), hold promise for achieving remarkable imaging quality with low sample rates. These methods employ GANs, or convolutional neural networks (CNNs), to enhance image reconstruction quality.
However, integrating DL into SPI presents its own set of challenges. DL-based techniques necessitate extensive training data and do not guarantee optimal image quality in unfamiliar scenes. To overcome this, researchers have explored the incorporation of untrained neural networks into physical models. Methods such as variable generative networks (VGenNet) and untrained reconstruction networks (URNet) utilize untrained neural networks to achieve high-quality imaging without the need for extensive training data.
Study Methodology
The SPI setup comprises an expanded laser source, a single-pixel detector (SPD), an object, a digital micromirror device (DMD), and a computer. The 10-mW laser with a 532 nm wavelength is expanded and modulated by the DMD using random speckle patterns stored in the computer. As the laser travels through the object, the SPD collects intensity signals, which are then transmitted to the computer for analysis.
ILNet leverages a part-based model for fine-grained learning, enhancing image details during reconstruction. The calculated intensity signals of the reconstructed image serve as a loss function for training ILNet's parameters. The refined, reconstructed image can undergo iterative improvement, and the 2D image produced by ILNet can be utilized as input for successive iterations. This imparts prior information to enhance imaging quality when dealing with low sampling rates.
ILNet's encoding-decoding structure employs convolutional and deconvolutional layers for feature extraction and restoration, respectively. Experimental setups encompass free-space and underwater environments, thoroughly assessing ILNet's effectiveness. Five laser-cut alphabet letters serve as objects, and quantitative evaluations are conducted using contrast-to-noise ratio (CNR), peak signal-to-noise ratio (PSNR), and resolution measures.
Major Findings
Initial comparisons are drawn between ILNet and conventional image reconstruction techniques such as traditional ghost imaging (TGI) and differential ghost imaging (DGI) within an unfamiliar free-space setting. Impressively, ILNet exhibits superior noise reduction and object reconstruction, even with fewer iterations. The employed part-based model within ILNet augments image intricacies, yielding higher-quality images with reduced iterations compared to TGI and DGI.
Quantitative assessments validate ILNet's superiority, showcasing heightened contrast-to-noise ratio (CNR) and peak signal-to-noise ratio (PSNR), even with fewer iterations. ILNet-generated image resolution significantly surpasses that of TGI and DGI.
Furthermore, investigations into diverse part configurations within ILNet are conducted. Results suggest that partitioning both the encoder and decoder leads to incomplete images, adversely affecting CNR, PSNR, and resolution. However, isolating partitioning solely to the encoder results in remarkably clear images, indicating adept capture of fine-grained features and superior performance.
ILNet's efficacy is then evaluated through a turbulent water-based reflected SPI experiment. It outperforms TGI, DGI, and Differential Ghost Imaging with Enhanced Positive Signals (SDGI) in terms of CNR, PSNR, and resolution. Unlike SDGI's binarization-centric approach that results in diminished CNR and PSNR due to binary pixel values, ILNet achieves enhanced image quality without relying solely on binarization. ILNet's incorporation of detected single-pixel values, part-based models, and image looping contributes to continuous optimization and improved overall image quality.
It is important to acknowledge that images retrieved in a turbulent water environment manifest distortions and omissions due to light scattering and refraction, challenges not entirely resolved by ILNet. Nevertheless, ILNet effectively achieves its primary goal of reducing sample rates and enhancing image quality.
Conclusion
In summary, the current study introduced ILNet-enhanced SPI by integrating physical models into a part-based neural network. This approach employs the model for refined 2D object image reconstruction. The iterative generation of 2D images continually enhances quality by incorporating prior information. By utilizing 1D SPD signals for adaptive optimization and object image reconstruction, ILNet excels in unknown environments, elevating image quality and stability. This enables high-quality SPI at extremely low sample rates, extending applications to optically harsh conditions.