Get 20M+ Full-Text Papers For Less Than $1.50/day. Start a 14-Day Trial for You or Your Team.

Learn More →

Chromatic-Aberration-Corrected Hyperspectral Single-Pixel Imaging

Chromatic-Aberration-Corrected Hyperspectral Single-Pixel Imaging hv photonics Article Chromatic-Aberration-Corrected Hyperspectral Single-Pixel Imaging 1 1 , 2 , 3 1 1 Ying Liu , Zhao-Hua Yang * , Yuan-Jin Yu * , Ling-An Wu , Ming-Yue Song and Zhi-Hao Zhao School of Instrumentation and Optoelectronics Engineering, Beihang University, Beijing 100191, China School of Automation, Beijing Institute of Technology, Beijing 100081, China Institute of Physics, Chinese Academy of Sciences, Beijing 100190, China * Correspondence: yangzh@buaa.edu.cn (Z.-H.Y.); yuanjin.yu@bit.edu.cn (Y.-J.Y.) Abstract: With the emerging development of hyperspectral single-pixel imaging (SPI) systems, the trade-off between the simplicity of optical structure and the correction of chromatic aberration is now an essential factor to be considered. To address both issues simultaneously, we propose a chromatic-aberration-corrected hyperspectral single-pixel imaging scheme, which is based on spectral measurement and dispersion correction. Its achromatism feature is evaluated by optical simulations and proof-of-concept experiments. Moreover, to overcome the shortcomings of traditional algorithms, a new adaptive iterative algorithm is employed, which can further optimize image quality. The results demonstrate that both dispersion and noise in our system are significantly reduced. Taking the position coordinate variance as a figure of merit, we have realized an order of magnitude improvement in the lateral chromatic aberration over the spectral range of 400–780 nm compared to that in conventional hyperspectral SPI. Meanwhile, the contrast-to-noise ratio in our system is enhanced on average by 3 dB. To the best of our knowledge, this is the first such demonstration, and the technique presents possibilities for future integrated applications of high spatial/spectral resolution over the entire visible range, and the system has the potential to be scaled down for future integrated applications. Keywords: single-pixel imaging; hyperspectral; chromatic aberration correction; optical system 1. Introduction Citation: Liu, Y.; Yang, Z.-H.; Yu, Y.-J.; The first recorded attempt at single-pixel imaging may be traced back to 1884, when Wu, L.-A.; Song, M.-Y.; Zhao, Z.-H. Paul Nipkow patented the flying-spot camera, using a Nipkow disk to modulate the Chromatic-Aberration-Corrected light source [1]. After years of development, contemporary single-pixel imaging (SPI) has Hyperspectral Single-Pixel Imaging. Photonics 2023, 10, 7. https:// advanced in parallel with computational ghost imaging (GI) [2,3]. The basic idea is to doi.org/10.3390/photonics10010007 modulate the reflected/transmitted beam illuminating an object with a sequence of patterns and record the corresponding total intensity values with a single-pixel (bucket) detector, Received: 28 October 2022 then recover the image through convolution. The patterns are generally pre-ordered and Revised: 18 December 2022 generated by a programmable spatial light modulator (SLM). The measured data, combined Accepted: 19 December 2022 with the prior information of the illumination speckles, constitute the essential elements of Published: 22 December 2022 a given computational reconstruction algorithm. Currently, there is a variety of reconstruction algorithms. Ferri et al. first proposed differential GI in 2010 [4], which has been widely adopted for its intrinsic advantage of anti- Copyright: © 2022 by the authors. noise interference. Later, many other algorithms were proposed, for instance, normalized Licensee MDPI, Basel, Switzerland. GI [5], high-order GI [6], correspondence GI [7] and self-evolving GI [8]. This article is an open access article In single-pixel imaging, only the total intensity values need to be measured, which distributed under the terms and makes it particularly suitable for applications in low visibility conditions, such as in remote conditions of the Creative Commons sensing [9,10] and underwater imaging [11]. Formerly, single-pixel sensing operated on Attribution (CC BY) license (https:// a single wavelength. When the process is extended to a wide spectral range, chromatic creativecommons.org/licenses/by/ aberration occurs since the same material exhibits different refractive indices at different 4.0/). Photonics 2023, 10, 7. https://doi.org/10.3390/photonics10010007 https://www.mdpi.com/journal/photonics Photonics 2023, 10, 7 2 of 12 wavelengths. This type of deviation would restrict the further enhancement of the spatial resolution of a broadband single-pixel imager. More severely, the larger the spectral range, the larger the chromatic aberration would be. Currently, known solutions, including temporal multiplexing [12] and frequency-division multiplexing [13], which are akin to introducing optical bandpass filters, predominantly suffer from the redundancy of optical structure and restricted spectral resolution. Schemes of coded aperture snapshot spectral imagers [14–16] could solve the problem, but with a significantly increased complexity of the algorithms. Hyperspectral imaging has been implemented based on SPI with a flat-field grating [17–19]. However, these schemes paid little attention to chromatic aberration since they were basically proximity range detection. To minimize the impact of chromatic aberration and to reduce structural vulnerability, spectral anti-aliasing with as simple as possible optical arrangements is required. In traditional spectral systems, this is achieved by dispersing the incident polychromatic light before the imaging section with an independent set of devices for each waveband, but these systems are always heavy and cumbersome. In this article, we propose a solution based on spectral measurement and dispersion correction, which we call chromatic-aberration-corrected hyperspectral single-pixel imaging (CHSPI). A spectrometer is used as the single-pixel sensor, so as to measure the spectrum of the reflected light. A closed-circuit television camera lens (later referred to as the TV lens) with a double Gaussian structure is adopted instead of a single thin lens. To eliminate the ripple noise when the object is imaged in a real light field and reconstructed by a traditional SPI algorithm, an adaptive and iterative correction algorithm is formulated. Simulations and a series of proof-of-concept experiments are conducted to test the achromatic capability of our scheme. As far as we are aware, this is the first demonstration of chromatic aberration correction in broadband hyperspectral SPI. The lateral chromatic aberration is suppressed by an order of magnitude over 400–780 nm in terms of pixel coordinate variance. Compared with previous schemes, our system can achieve high-contrast achromatic hyperspectral imaging over a broad range while the structure is lightweight and stable—characteristics that are particularly desirable in remote sensing applications. 2. System Model and Theoretical Derivation 2.1. Principle of Hyperspectral Differential Ghost Imaging In our scheme, a differential GI-based algorithm was selected to reconstruct the image while suppressing the noise [4]. In this linear non-iterative method, a variable approximate to the average intensity of the light field is subtracted from each pixel. In this way, the effective signal of the bucket detector just reflects the fluctuations of the light so that the noise is significantly suppressed and the signal-to-noise ratio (SNR) of the reconstructed Photonics 2022, 9, x FOR PEER REVIEW 3 of 14 image is improved. The basic principle of SPI is shown in Figure 1, where SLM is a spatial light modulator. Figure 1. Principle of single-pixel imaging. Figure 1. Principle of single-pixel imaging. In traditional SPI, the image is restored by correlating the light field modulation func- tion I(x) with the bucket signal S: (1) SI = ()xxR()dx where R(x) is the (intensity) reflection function of the object, and is the plane on which the image is modulated. The retrieved image is thus: G()xx = δδSI ⋅ () (2) 1 N δII ()xx=− () I(x) where denotes the ensemble average and . In differ- ... = (...) i =1 δRR ()xx=− () R ential GI, the fluctuating part of R(x) is defined as , where R=( I()xx Rxx)d I( )d x  is the average transmission function of the object. We de- fine a new differential bucket signal as: SI = ()xx δR()dx (3) Suppose NN × is the number of pixels in each pattern and K is the number of pre- set 2D spatial patterns (KN ≤×N ). Since the shot noise is negligible compared with the NN ××K abundant detected photons, the known modulation matrix M x ∈ and the beam () NN ××K I x ∈ , which is reflected by the SLM, are perfectly correlated variables; we can () IM ()xx = α ( ) introduce a factor to describe the correlation as . At the same time, we SM = ()xxd can set a virtual bucket signal to simplify the expression. Notice that can be expressed in terms of as: R = (4) α S In this way, S1 can be measured from the following operative form SS =− S (5) 1 R The image is then recovered by substituting S with S1 in Equation (2), namely, Photonics 2023, 10, 7 3 of 12 In traditional SPI, the image is restored by correlating the light field modulation function I(x) with the bucket signal S: S = I(x)R(x)dx (1) where R(x) is the (intensity) reflection function of the object, and A is the plane on which the image is modulated. The retrieved image is thus: G(x) = dS dI(x) (2) h i where . . . = (. . .) denotes the ensemble average and dI(x) = I(x) I(x) . h i å h i N i=1 In differential GI, the fluctuating part of R(x) is defined as dR(x) = R(x) R, where R R 2 2 R = hI(x)iR(x)d x/ hI(x)id x is the average transmission function of the object. We define a new differential bucket signal S as: S = I(x)dR(x)dx (3) Suppose N N is the number of pixels in each pattern and K is the number of preset 2D spatial patterns (K  N N). Since the shot noise is negligible compared with the NNK abundant detected photons, the known modulation matrix M(x) 2 R and the beam NNK I(x) 2 R , which is reflected by the SLM, are perfectly correlated variables; we can introduce a factor a to describe the correlation as I(x) = aM(x). At the same time, we can set a virtual bucket signal S = M(x)d x to simplify the expression. Notice that R can be expressed in terms of hS i as: 1 hSi R = (4) a S h i In this way, S can be measured from the following operative form h i S = S S (5) 1 R hS i The image is then recovered by substituting S with S in Equation (2), namely, G (x) = dS  dI(x) h i 1 1 (6) hSi = hS I(x)i hS I(x)i hS i In hyperspectral imaging, the above reconstruction of grayscale images corresponds to periodic sampling in the spectral dimension. More specifically, the detected intensity values over every 10 nm of the spectrum are summed as the “bucket detector value” S of the corresponding band I, from which the grayscale image G (x) is reconstructed as in 1i Equation (7) below: G (x) = hdS dI(x)i i 1i (7) hS i = hS  I(x)i hS  I(x)i i i hS i iR Based on the CIE 1931 Standard Colorimetric Observer, within the visible spectrum of 400–780 nm, the weighted sum of the images in each band is used to obtain the mixed color tristimulus image X(x), Y(x) and Z(x): X(x) = k G (x) x(l)Dl (8) å 1i Y(x) = k G (x) y(l)Dl (9) å 1i Z(x) = k G (x) z(l)Dl (10) å 1i where x(l), y(l), and z(l) are the color matching functions, k is the adjustment coefficient, and Dl is a fixed wavelength interval [20]. Then, by respectively plugging the three Photonics 2022, 9, x FOR PEER REVIEW 4 of 14 G()xx = δδSI ⋅ () 1 1 (6) =⋅SII ()xx − S⋅ ( ) In hyperspectral imaging, the above reconstruction of grayscale images corresponds to periodic sampling in the spectral dimension. More specifically, the detected intensity values over every 10 nm of the spectrum are summed as the “bucket detector value” G () x 1i of the corresponding band I, from which the grayscale image is reconstructed as in Equation (7) below: G (x) = δδ S ⋅ I() x 1i 1i S (7) =⋅ SII ()xx − S⋅ () ii iR Based on the CIE 1931 Standard Colorimetric Observer, within the visible spectrum of 400‒780 nm, the weighted sum of the images in each band is used to obtain the mixed color tristimulus image X(x), Y(x) and Z(x): XG ()xk=⋅ ()xx(λλ)Δ (8)  1i YG ()xk=⋅ ()xy(λλ)Δ (9) 1i ZG ()xk=⋅ ()xz(λλ)Δ (10)  1i Photonics 2023, 10, 7 4 of 12 where x() λ , y() λ , and z() λ are the color matching functions, k is the adjustment coef- ficient, and Δλ is a fixed wavelength interval [20]. Then, by respectively plugging the three matrices X(x), Y(x), and Z(x) into the three channels of the resultant RGB image, the matrices X(x), Y(x), and Z(x) into the three channels of the resultant RGB image, the entire entire reconstruction process can be completed. reconstruction process can be completed. 2.2. Principle of Chromatic Aberration Correction 2.2. Principle of Chromatic Aberration Correction The refractive index of a given material varies with the wavelength of light. Therefore, The refractive index of a given material varies with the wavelength of light. There- in a typical single thin lens imaging system, the object is imaged at different locations fore, in a typical single thin lens imaging system, the object is imaged at different locations when illuminated by polychromatic light; in other words, the image would be aliased when illuminated by polychromatic light; in other words, the image would be aliased into into defocused speckles at any position [20], as shown in Figure 2. Thus, this chromatic defocused speckles at any position [20], as shown in Figure 2. Thus, this chromatic aber- aberration would decrease the accuracy of R(x) in Equation (1). ration would decrease the accuracy of R(x) in Equation (1). Figure 2. Schematic of chromatic aberration along the optical axis. A: object beam. Wavelengths of F, D, and C are 486 nm, 588 nm, and 656 nm, respectively. To mitigate the problem, achromatic design is required. For a visual system, D light (588 nm) is generally chosen as the reference wavelength. The imaging positions of F light (486 nm) and C light (656 nm) are designed to coincide with the reference. The deviation between the imaging positions of two colors on the optical axis is called positional chromatic aberration, represented by 2 i 0 0 0 DL = L L = (11) FC F C 2 2 n0 u0 0 0 where h is the projection height of the object on each lens, n and u are the refractive index and aperture angle on the image side, and j and n are the focal power and Abbe coefficient i i of each lens, respectively [20]. To eliminate the chromatic aberration, h is expected to be zero. It can be seen that a single lens cannot correct the chromatic dispersion because its focal power, j, is always a non-zero constant but may be negative or positive depending on whether the lens 2 i is convex or concave, respectively. By selecting a suitable combination of lenses, h can be reduced to zero. Our scheme uses broadband light to illuminate a linear array detector for detection, so eliminating chromatic aberration is a key issue. Meanwhile, considering the advantage of reducing vertical-axis aberration (such as distortion, etc.), the double Gaussian lens is used. The double Gaussian lens is a typical successful adoption of the above combination, as shown in Figure 3. Two meniscus lenses, one positive meniscus L and another negative meniscus L , can reduce the spherical aberration and longitudinal chromatic aberration by 2 Photonics 2022, 9, x FOR PEER REVIEW 5 of 14 Figure 2. Schematic of chromatic aberration along the optical axis. A: object beam. Wavelengths of F, D, and C are 486 nm, 588 nm, and 656 nm, respectively. To mitigate the problem, achromatic design is required. For a visual system, D light (588 nm) is generally chosen as the reference wavelength. The imaging positions of F light (486 nm) and C light (656 nm) are designed to coincide with the reference. The deviation between the imaging positions of two colors on the optical axis is called positional chro- matic aberration, represented by ν (11) Δ= L' L' − L' =− FC F C n' u' where h is the projection height of the object on each lens, n' and u' are the refractive index and aperture angle on the image side, and ϕ and ν are the focal power and i i Abbe coefficient of each lens, respectively [20]. 2 i To eliminate the chromatic aberration, h is expected to be zero. It can be seen that a single lens cannot correct the chromatic dispersion because its focal power, ϕ , is always a non-zero constant but may be negative or positive depending on whether the lens is convex or concave, respectively. By selecting a suitable combination of lenses, h can be reduced to zero. Our scheme uses broadband light to illuminate a linear array detector for detection, so eliminating chromatic aberration is a key issue. Meanwhile, considering the advantage of reducing vertical-axis aberration (such as distortion, etc.), the double Gaussian lens is used. The double Gaussian lens is a typical successful adoption of the above combination, Photonics 2023, 10, 7 5 of 12 as shown in Figure 3. Two meniscus lenses, one positive meniscus L1 and another negative meniscus L2, can reduce the spherical aberration and longitudinal chromatic aberration by splitting one positive lens into two so that the curvature is spread over four surfaces splitting one positive lens into two so that the curvature is spread over four surfaces instead instead of two, while the overall power of the lens is preserved. Subsequently, placing of two, while the overall power of the lens is preserved. Subsequently, placing lenses L lenses L3 and L4 to face L1 and L2 symmetrically would notably diminish the vertical axis and L to face L and L symmetrically would notably diminish the vertical axis aberration, 4 1 2 aberration, such as distortion. The vital step to correct the chromatic aberration is through such as distortion. The vital step to correct the chromatic aberration is through replacing replacing the negative meniscus lenses L2 and L3 with hyperchromatic lenses L21, L22 and the negative meniscus lenses L and L with hyperchromatic lenses L , L and L , L , 2 3 21 22 31 32 L31, L32, respectively. A hyperchromatic lens is a cemented doublet lens with the same respectively. A hyperchromatic lens is a cemented doublet lens with the same refractive refractive index at the given wavelength but different dispersion, such that changing the index at the given wavelength but different dispersion, such that changing the radius of the radius of the curvature of the cemented surface does not alter the focal length of the pri- curvature of the cemented surface does not alter the focal length of the primary wavelength mary wavelength but adjusts the chromatic aberration [21]. By arranging the above lenses but adjusts the chromatic aberration [21]. By arranging the above lenses in proper positions, in proper positions, the final double Gaussian lens can correct chromatic aberration. the final double Gaussian lens can correct chromatic aberration. Photonics 2022, 9, x FOR PEER REVIEW 6 of 14 Figure 3. Double Gaussian structure. Figure 3. Double Gaussian structure. Figure 3. Double Gaussian structure [21]. 3. Simulation 3. Simulation A series of simulations were performed to compare the capability to correct for chro- A series of simulations were performed to compare the capability to correct for chro- matic aberration in conventional hyperspectral SPI (below referred to as HSPI) and our matic aberration in conventional hyperspectral SPI (below referred to as HSPI) and our CHSPI systems. The focal length and aperture of the two groups were equally set to 50 mm CHSPI systems. The focal length and aperture of the two groups were equally set to 50 and 25 mm, respectively. Figure 4 shows the general layouts, the lateral chromatic aberra- mm and 25 mm, respectively. Figure 4 shows the general layouts, the lateral chromatic tion curves, and the chromatic focal shift curves of the two systems. The lateral chromatic aberration curves, and the chromatic focal shift curves of the two systems. The lateral aberration is generally measured by the furthest deviation from the vertical axis in the field chromatic aberration is generally measured by the furthest deviation from the vertical axis and the chromatic focal shift by the maximum difference between the focal positions in the in the field and the chromatic focal shift by the maximum difference between the focal given wavelength range. As the diagrams in Figure 4c,d illustrate, the maximal absolute positions in the given wavelength range. As the diagrams in Figure 4c,d illustrate, the value of lateral chromatic aberration in the CHSPI system is only 1/7 of that in the HSPI maximal absolute value of lateral chromatic aberration in the CHSPI system is only 1/7 of system over the wavelength range of 400–780 nm and the corresponding chromatic focal that in the HSPI system over the wavelength range of 400‒780 nm and the corresponding shift 1/15. chromatic focal shift 1/15. Figure 4. Simulation results of a single-lens system and TV lens system. Layout of imaging parts in Figure 4. Simulation results of a single-lens system and TV lens system. Layout of imaging parts (a) hyperspectral SPI (HSPI) and (b) chromatic-aberration-corrected hyperspectral single-pixel im- in (a) hyperspectral SPI (HSPI) and (b) chromatic-aberration-corrected hyperspectral single-pixel aging (CHSPI). (c) Lateral chromatic aberration curve: field of view vs. lateral chromatic aberration. imaging (CHSPI). (c) Lateral chromatic aberration curve: field of view vs. lateral chromatic aberration. (d) The chromatic focal shift curve: wavelength vs. chromatic focal shift in the range from 486 nm of F-light to 656 nm of C-light. (d) The chromatic focal shift curve: wavelength vs. chromatic focal shift in the range from 486 nm of F-light to 656 nm of C-light. 4. Experimental Details To verify the effectiveness of dispersion correction of our hyperspectral correlation imaging experimental system, we conducted an experiment consisting of a system based on the TV lens and a control group based on a single lens. The dispersion correction per- formance of the two systems was evaluated by comparing the relative deviations of the reconstructed images of different spectral bands at corresponding points in the recovered single-pixel images of the two groups under the same environmental conditions. The schematic diagram and actual view of our experimental setup are illustrated in Figure 5. A broadband beam of light from a supercontinuum laser source (LEUKOS ROCK 400, 400–2400 nm) was expanded to illuminate the object, a 3 × 3 × 3 Rubik’s cube of size 54 × 54 × 54 mm , which was then imaged by a TV lens onto a common type of SLM—a digital micromirror device (DMD) (ViALUX V-9501, 1920 × 1080) prepro- grammed with a series of Hadamard matrices. After modulation, the beam was focused into a spectrometer (Ocean Optics FLAME-S-UV-VIS, 340–1100 nm). We employed 4096 speckle patterns generated under the cake-cutting basis Hadamard basis algorithm [22] for both the control and TV lens measurements. Photonics 2023, 10, 7 6 of 12 4. Experimental Details To verify the effectiveness of dispersion correction of our hyperspectral correlation imaging experimental system, we conducted an experiment consisting of a system based on the TV lens and a control group based on a single lens. The dispersion correction performance of the two systems was evaluated by comparing the relative deviations of the reconstructed images of different spectral bands at corresponding points in the recovered single-pixel images of the two groups under the same environmental conditions. The schematic diagram and actual view of our experimental setup are illustrated in Figure 5. A broadband beam of light from a supercontinuum laser source (LEUKOS ROCK 400, 400–2400 nm) was expanded to illuminate the object, a 3  3  3 Rubik’s cube of size 54  54  54 mm , which was then imaged by a TV lens onto a common type of SLM—a digital micromirror device (DMD) (ViALUX V-9501, 1920 1080) preprogrammed with a series of Hadamard matrices. After modulation, the beam was focused into a spectrometer (Ocean Optics FLAME-S-UV-VIS, 340–1100 nm). We employed 4096 speckle Photonics 2022, 9, x FOR PEER REVIEW 7 of 14 patterns generated under the cake-cutting basis Hadamard basis algorithm [22] for both the control and TV lens measurements. Figure 5. Schematic (a) and photograph (b) of the CHSPI setup. Figure 5. Schematic (a) and photograph (b) of the CHSPI setup. 5. Algorithm Optimization 5. Algorithm Optimization In the process of reconstructing the grayscale image of each band, erroneous black In the process of reconstructing the grayscale image of each band, erroneous black points (zero points) are prone to appear at the upper-left edge, forming noise ripples, as points (zero points) are prone to appear at the upper-left edge, forming noise ripples, as shown in Figure 6a. For a picture with a homogeneous background and a contrasting shown in Figure 6a. For a picture with a homogeneous background and a contrasting ob- object, these segments should have been part of and not darker than the background. As ject, these segments should have been part of and not darker than the background. As Equation (2) shows, the essence of GI computation is matrix multiplication. Under the Equation (2) shows, the essence of GI computation is matrix multiplication. Under the full-sampling condition, the solution should be unique. Because the modulation matrix is a full-sampling condition, the solution should be unique. Because the modulation matrix is constant known parameter, only various unpredictable noises can explain this error. To a constant known parameter, only various unpredictable noises can explain this error. To minimize these errors while retaining the original information as much as possible, several minimize these errors while retaining the original information as much as possible, sev- methods are adopted. According to the ripple characteristics, the pixels adjacent to the eral methods are adopted. According to the ripple characteristics, the pixels adjacent to error point are also prone to error, which excludes the traditional median or mean filtering the error point are also prone to error, which excludes the traditional median or mean methods of correction. Nor can frequency domain filtering be used since the ripples do not filtering methods of correction. Nor can frequency domain filtering be used since the rip- appear in all areas of the image. Therefore, on the basis of differential GI but to overcome ples do not appear in all areas of the image. Therefore, on the basis of differential GI but its inadequacies, we designed a new adaptive iterative correction algorithm, as illustrated to overcome its inadequacies, we designed a new adaptive iterative correction algorithm, in Figure 6b. as illustrated in Figure 6b. In each iteration, the gray value of the error point is set as the global gray scale average to eliminate its impact on the image contrast before re-normalization. The error point i i always preferentially appears at the position of (1, 2 ) or (2 , 1)(i = 1, 2, 3 . . . . . . ). After i i1 i i1 each correction, its position moves from (1, 2 ) to (1, 2 + 1) or from (2 , 1) to (2 + 1, 1) but never beyond the upper-left quarter of the image. We define the edge region as two line segments from (1, 2) to (1, N/2) and from (2, 1) to (N/2, 1) to simplify the computation. Each time, we search for the first zero point from left to right and from top to bottom. The whole process is finished when the first zero point no longer appears in the edge region. The SPI images of different spectral bands before and after correction are shown in Figure 6c. To quantitatively evaluate the image quality, we use the contrast-to-noise ratio Photonics 2023, 10, 7 7 of 12 (CNR) instead of the conventional signal-to-noise ratio because we do not have an original object image for comparison. The CNR is defined as [23]: CNR =10 log (S /s ) (12) where S is the variance of the object area, and s is the variance of the background area. o b An average increase of more than 3 dB in Figure 6c clearly exemplifies the improvement in Photonics 2022, 9, x FOR PEER REVIEW 8 of 14 image quality by our algorithm. Figure 6. Proposed adaptive iterative correction algorithm. (a) The noise ripples (zero points) on the Figure 6. Proposed adaptive iterative correction algorithm. (a) The noise ripples (zero points) on the upper-left edge. (b) The adaptive iterative correction algorithm used to correct the erroneous black upper-left edge. (b) The adaptive iterative correction algorithm used to correct the erroneous black noise ripples. (c) Comparison of SPI images in different spectral bands before and after correction. noise ripples. (c) Comparison of SPI images in different spectral bands before and after correction. In each iteration, the gray value of the error point is set as the global gray scale aver- 6. Results and Discussion age to eliminate its impact on the image contrast before re-normalization. The error point i i always preferentially appears at the position of (1, 2 ) or (2 , 1)(i = 1, 2, 3……). After each The optimized results of each wavelength band of hyperspectral SPI in the control i i−1 i i−1 correction, its position moves from (1, 2 ) to (1, 2 + 1) or from (2 , 1) to (2 + 1, 1) but never and experimental groups are shown in Figure 7, where pseudo-coloring has been applied. beyond the upper-left quarter of the image. We define the edge region as two line seg- In Figure 8, under the CIE1931 standard observer model, images of all spectral bands are ments from (1, 2) to (1, N/2) and from (2, 1) to (N/2, 1) to simplify the computation. Each fused to obtain the reconstructed color images. In the control group of Figure 8a, we can time, we search for the first zero point from left to right and from top to bottom. The whole see the influence of lateral chromatic aberration, as images are only clear from 580–610 nm, process is finished when the first zero point no longer appears in the edge region. The SPI elsewhere being blurred to varying degrees. By contrast, in the aberration-corrected images images of different spectral bands before and after correction are shown in Figure 6c. To in Figure 8b, the block margins are sharply outlined and distinct from the background quantitatively evaluate the image quality, we use the contrast-to-noise ratio (CNR) instead for all the bands; in addition, we can even discern the patchy reflections from the uneven of the conventional signal-to-noise ratio because we do not have an original object image surfaces of the colored squares, which are actually composed of glossy paper glued onto a for comparison. The CNR is defined as [23]: black cube. CNR=10log () S / σ (12) 10 ob where S is the variance of the object area, and σ is the variance of the background o b area. An average increase of more than 3 dB in Figure 6c clearly exemplifies the improve- ment in image quality by our algorithm. Photonics 2023, 10, 7 8 of 12 Photonics 2022, 9, x FOR PEER REVIEW 10 of 14 Figure 7. Reconstructed hyperspectral images over 400‒780 nm in the control (a) and TV lens (b) Figure 7. Reconstructed hyperspectral images over 400–780 nm in the control (a) and TV lens (b) groups, with their composite-colored images in the last figure. (c,d): Enlargements of the respective groups, with their composite-colored images in the last figure. (c,d): Enlargements of the respective images in the red boxes of (a,b) reveal that sharp edges are only seen in the yellow bands of the images single-lens in the syst red em, while boxes they of (a ,are di b) reveal stinctthat oversharp the vast edges majority of band are only seen range of the TV in the yellow system. bands of the single-lens system, while they are distinct over the vast majority of band range of the TV system. Photonics 2023, 10, 7 9 of 12 Photonics 2022, 9, x FOR PEER REVIEW 11 of 14 Figure 8. Reconstructed colored images of the control group (a) and the TV lens group (b). The white Figure 8. Reconstructed colored images of the control group (a) and the TV lens group (b). The white boxes show the main differences: purple or yellow haloes appear at the edges of the color blocks in boxes show the main differences: purple or yellow haloes appear at the edges of the color blocks in (a), while in (b), even details such as the white edges in boxes B2 and C2 are clearly revealed. (a), while in (b), even details such as the white edges in boxes B2 and C2 are clearly revealed. In order to quantify the performance of the two schemes, we examine a cross-section profile in the two resultant images (shown by the red lines in Figure 9a,c). The pixels’ gray values on the red line in each spectral band of 400–780 nm are successively extracted from left to right and plotted in the form of line charts in Figure 9b,d. Among them, six spectral bands (440 nm for blue, 490 nm for cyan, 530 nm for green, 580 nm for yellow, 630 nm for orange, and 680 nm for red) are selected. In the two sets of figures, three groups of pixels 1 , 2 , and 3 at the object’s edge positions are chosen where there are marked discontinuous changes in the gray values. Group corresponds to the edge of the white block and the maximum gray value point in the region, while group 2 in the center of the black gap corresponds to the minimum gray value. Group in the TV lens group of Figure 9c corresponds to the reflected white edge and the minimum gray value in the region, but in the control group of Figure 9a, this detail is too blurred, so the edge of the orange block is chosen, which corresponds to the maximum gray value in this region. In Figure 9b,d, each group is labelled and connected across the spectral bands with a blue line, while the vertical red dotted lines denote the abscissa of the chosen points with the first band (440 nm) as the reference origin. As the wavelength increases in Figure 9b, the corresponding points in the same group move to the left due to lateral chromatism. Additionally, the distance between groups 1 and 2 is significantly different in the change from blue to red; that is, from 3 pixels apart in the blue band to only 1 pixel apart in the red band, which also indicates the existence of lateral chromatism. However, in Figure 9d, we can see that the blue lines connecting the groups are straighter and almost coincide with the red lines, showing that aberration is corrected and variation with wavelength diminished. The overall deviations of the three positions from the reference origin point, as a function of wavelength, are shown in Figure 10a,b. In order to quantify the chromatic aberration, we calculated the pixel column number variances of the deviation at the three positions, as shown in Table 1. It can be seen that the Figure 9. Comparison of the performance of chromatic aberration correction. The intensity profiles variances in the TV lens scheme are on average more than an order of magnitude smaller across the red lines in the reconstructed images of (a) the control group and (c) the TV lens group than those in the single-lens scheme. vs. pixel number for different wave bands are shown in (b,d). Three groups of pixels ①, ②, and ③ are selected and in (b,d) are connected across the spectral bands with a blue line, while the vertical red dotted lines denote the abscissa of the chosen points with the first band (440 nm) as the reference origin. Photonics 2022, 9, x FOR PEER REVIEW 11 of 14 Figure 8. Reconstructed colored images of the control group (a) and the TV lens group (b). The white Photonics 2023, 10, 7 10 of 12 boxes show the main differences: purple or yellow haloes appear at the edges of the color blocks in (a), while in (b), even details such as the white edges in boxes B2 and C2 are clearly revealed. Figure 9. Comparison of the performance of chromatic aberration correction. The intensity profiles Figure 9. Comparison of the performance of chromatic aberration correction. The intensity profiles acr across the red lines in the reco oss the red lines in the reconstr nstructed images of ( ucted images of (a a)) the co the contr ntrol group and ( ol group and (c c) the TV le ) the TV lens ns group group vs. pixel number for different wave bands are shown in (b,d). Three groups of pixels ①, ②, and ③ vs. pixel number for different wave bands are shown in (b,d). Three groups of pixels 1 , 2 , and are selected and in (b,d) are connected across the spectral bands with a blue line, while the vertical are selected and in (b,d) are connected across the spectral bands with a blue line, while the Photonics 2022, 9, x FOR PEER REVIEW 12 of 14 red dotted lines denote the abscissa of the chosen points with the first band (440 nm) as the reference vertical red dotted lines denote the abscissa of the chosen points with the first band (440 nm) as the origin. reference origin. Figure 10. Pixel column number of the ①, ②, and ③ positions vs. wavelength over 400–780 nm for Figure 10. Pixel column number of the 1 , 2 , and 3 positions vs. wavelength over 400–780 nm for the control group (a) and the TV lens group (b). A flat line indicates no deviation in the focal point the control group (a) and the TV lens group (b). A flat line indicates no deviation in the focal point position, i.e. no lateral chromatic aberration occurs. position, i.e. no lateral chromatic aberration occurs. In order to quantify the chromatic aberration, we calculated the pixel column number variances of the deviation at the three positions, as shown in Table 1. It can be seen that the variances in the TV lens scheme are on average more than an order of magnitude smaller than those in the single-lens scheme. Table 1. Position Coordinate Variance in the Two Schemes. Scheme Name ① ② ③ HSPI scheme 4.8889 7.1137 1.8659 CHSPI scheme 0.1473 0.2459 0.2025 Improvement (HSPI/CHSPI) 33.1901 28.9292 9.2143 The theoretical spatial resolution limits of conventional imaging and the achieved resolutions of the two SPI systems are compared below to quantify their performance from another perspective. The apertures DC and DT of the single lens and TV lens are, respectively, 30 mm and 16 mm. The smallest unit that was clearly resolved by the single lens system was the 4 mm width of the black gaps between the colored blocks in A1 of Figure 8a, whereas the TV lens system could clearly resolve the white reflective strips in B2 and C2 of Figure 8b, which have a width of less than 1 mm. The distance between the object and the lens is more than 45 cm. Thus, the theoretical and actual angular resolutions can be derived as follows. For a central wavelength λ of 550 nm the theoretical resolution limit of conventional imaging using a single-lens system is ϕλ =1.22 / D =× 1.22 550nm/30mm=0.022mrad (13) CC while that of the TV lens system is ϕλ =1.22 / D =× 1.22 550nm/16mm=0.042mrad (14) TT The spatial resolution actually achieved in the single-lens SPI setup was (15) ϕ ≈ΔlL / = 4mm/450mm =8.9mrad while that in the improved SPI system was (16) ϕ ≈ΔlL / =1mm/450mm = 2.2mrad Photonics 2023, 10, 7 11 of 12 Table 1. Position Coordinate Variance in the Two Schemes. 1 2 3 Scheme Name HSPI scheme 4.8889 7.1137 1.8659 CHSPI scheme 0.1473 0.2459 0.2025 Improvement (HSPI/CHSPI) 33.1901 28.9292 9.2143 The theoretical spatial resolution limits of conventional imaging and the achieved resolutions of the two SPI systems are compared below to quantify their performance from another perspective. The apertures D and D of the single lens and TV lens are, C T respectively, 30 mm and 16 mm. The smallest unit that was clearly resolved by the single lens system was the 4 mm width of the black gaps between the colored blocks in A1 of Figure 8a, whereas the TV lens system could clearly resolve the white reflective strips in B2 and C2 of Figure 8b, which have a width of less than 1 mm. The distance between the object and the lens is more than 45 cm. Thus, the theoretical and actual angular resolutions can be derived as follows. For a central wavelength l of 550 nm the theoretical resolution limit of conventional imaging using a single-lens system is j = 1.22l/D = 1.22 550 nm/30 mm = 0.022 mrad (13) C C while that of the TV lens system is j = 1.22l/D = 1.22 550 nm/16 mm = 0.042 mrad (14) T T The spatial resolution actually achieved in the single-lens SPI setup was j  Dl/L = 4 mm/450 mm = 8.9 mrad (15) while that in the improved SPI system was j  Dl/L = 1 mm/450 mm = 2.2 mrad (16) Above are merely the results of our recent proof-of-principle experiments at present, and the actual spatial resolution limits need to be explored in subsequent research. Al- though the current measured resolutions are far from the theoretical values, comparing Equations (15) and (16), we can see that the achieved spatial resolution of the optimized TV lens system is more than a factor of four better than that of the control setup, even though its aperture is much smaller. 7. Conclusions As the first reported demonstration of chromatic aberration correction in hyperspectral SPI, our scheme succeeded in achieving an order of magnitude improvement in image quality across a wide spectral range of 400–780 nm. The images are reconstructed via an im- proved adaptive and iterative correction single-pixel algorithm, which allows information across all bandwidths to be retrieved within a short period of time with an image quality enhancement of more than 3 dB on average. Our scheme could provide a technical solution for spectral de-aliasing in SPI settings for microscopy and general analysis while providing a compact, lightweight, stable solution for remote sensing and satellite applications. Author Contributions: Conceptualization, Y.L.; methodology, Y.L.; software, Y.L.; validation, Y.L., M.-Y.S., and Z.-H.Z.; investigation, Y.L.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, Z.-H.Y., Y.-J.Y., and L.-A.W.; visualization, Y.L.; funding acqui- sition, Z.-H.Y., Y.-J.Y., and L.-A.W. All authors have read and agreed to the published version of the manuscript. Photonics 2023, 10, 7 12 of 12 Funding: This work was supported in part by the National Natural Science Foundation of China under Grants 61973018, 62173039, and 61975229, the Civil Space Project under Grant D040301, the National Key Research and Development Program of China under Grant 2018YFB0504302, and the Defense Industrial Technology Development Program under grant JCKY2021602B036. Institutional Review Board Statement: Not applicable. Informed Consent Statement: Not applicable. Data Availability Statement: Not applicable. Conflicts of Interest: The authors declare no conflict of interest. References 1. Yin, S.; Lu, G.; Zhang, J.; Yu, F.; Mait, J.N. Kinoform-based nipkow disk for a confocal microscope. Appl. Opt. 1995, 34, 5695. [CrossRef] [PubMed] 2. Shapiro, J.H. Computational ghost imaging. Phys. Rev. A 2008, 78, 061802. [CrossRef] 3. Shapiro, J.H.; Erkmen, B.I.; Lvovsky, A. Ghost imaging: From quantum to classical to computational. Adv. Opt. Photonics 2010, 2, 405–450. 4. Ferri, F.; Magatti, D.; Lugiato, L.A.; Gatti, A. Differential ghost imaging. Phys. Rev. Lett. 2010, 104, 253603. [CrossRef] [PubMed] 5. Sun, B.Q.; Welsh, S.S.; Edgar, M.P.; Shapiro, J.H.; Padgett, M.J. Normalized ghost imaging. Opt. Express 2012, 20, 16892–16901. [CrossRef] 6. Chen, X.H.; Agafonov, I.N.; Luo, K.H.; Liu, Q.; Xian, R.; Chekhova, M.V.; Wu, L.-A. High-visibility, high-order lensless ghost imaging with thermal light. Opt. Lett. 2010, 35, 1166–1168. [CrossRef] 7. Luo, K.H.; Huang, B.Q.; Zheng, W.-M.; Wu, L.-A. Nonlocal Imaging by Conditional Averaging of Random Reference Measure- ments. Chin. Phys. Lett. 2012, 29, 074216. [CrossRef] 8. Liu, B.; Wang, F.; Chen, C.; Dong, F.; McGloin, D. Self-evolving ghost imaging. Optica 2021, 8, 1340–1349. [CrossRef] 9. Erkmen, B.I. Computational ghost imaging for remote sensing. J. Opt. Soc. Am. A Opt. Image Sci. Vis. 2012, 29, 782–789. [CrossRef] 10. Yang, H.; Huang, Y.-H.; Zhong, C.; Shao, X.-P. Impact of Satellite Movement on Compressive Coded Aperture Spectral Imaging. Spacecr. Recovery Remote Sens. 2013, 34, 16–24. 11. Le, M.; Gao, W.; Zheng, H.; Liu, J.; Zhuo, X. Underwater computational ghost imaging. Opt. Express 2017, 25, 22859. [CrossRef] 12. Wang, Y.; Suo, J.; Fan, J.; Dai, Q. Hyperspectral Computational Ghost Imaging via Temporal Multiplexing. IEEE Photonics Technol. Lett. 2016, 28, 288–291. [CrossRef] 13. Bian, L.; Suo, J.; Situ, G.; Li, Z.; Fan, J.; Chen, F.; Dai, Q. Multispectral imaging using a single bucket detector. Sci. Rep. 2016, 6, 24752. [CrossRef] [PubMed] 14. Wagadarikar, A.A.; John, R.; Willett, R.; Brady, D. Single disperser design for coded aperture snapshot spectral imaging. Appl. Optics 2008, 47, B44–B51. [CrossRef] [PubMed] 15. Arguello, H.; Arce, G. Code aperture design for compressive spectral imaging. In Proceedings of the 2010 European Signal Processing Conference (EUSIPCO-2010), Aalborg, Denmark, 23–27 August 2010. 16. Arguello, H.; Arce, G.R. Rank Minimization Code Aperture Design for Spectrally Selective Compressive Imaging. IEEE Trans. Image Process. 2013, 22, 941–954. [CrossRef] [PubMed] 17. Chao, W.; Liu, X.F.; Yu, W.K.; Yao, X.R.; Li, L.Z.; Zhao, Q.; Zhai, G.-J. Compressed spectral imaging with a spectrometer. Opt. Commun. 2015, 352, 45–48. 18. Liu, S.; Liu, Z.; Wu, J.; Li, E.; Hu, C.; Tong, Z.; Shen, X.; Han, S. Hyperspectral ghost imaging camera based on a flat-field grating. Opt. Express 2018, 26, 17705–17716. [CrossRef] [PubMed] 19. Liu, B.; Wang, F.; Chen, C.; McGloin, D. Single-Pixel Diffuser Camera. IEEE Photonics J. 2021, 13, 1–5. [CrossRef] 20. Yu, D.-Y.; Tan, H.-Y. Engineering Optics, 4th ed.; China Machina Press: Beijing, China, 2015; pp. 93–94, 126–129. 21. Mandler, W. Design of Basic Double Gauss Lenses. In Proceedings of the 1980 International Lens Design Conference, Oakland, America, 16 September 1980. 22. Yu, W.K. Super sub-Nyquist single-pixel imaging by means of cake-cutting Hadamard basis sort. Sensors 2019, 19, 4122. [CrossRef] [PubMed] 23. Li, M.-F.; Yan, L.; Yang, R.; Kou, J.; Liu, Y.-X. Turbulence-free intensity fluctuation self-correlation imaging with sunlight. Acta Phys. Sin. 2019, 68, 094204. [CrossRef] Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. http://www.deepdyve.com/assets/images/DeepDyve-Logo-lg.png Photonics Multidisciplinary Digital Publishing Institute

Chromatic-Aberration-Corrected Hyperspectral Single-Pixel Imaging

Loading next page...
 
/lp/multidisciplinary-digital-publishing-institute/chromatic-aberration-corrected-hyperspectral-single-pixel-imaging-da3uBxJhCx

References (24)

Publisher
Multidisciplinary Digital Publishing Institute
Copyright
© 1996-2022 MDPI (Basel, Switzerland) unless otherwise stated Disclaimer Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. Terms and Conditions Privacy Policy
ISSN
2304-6732
DOI
10.3390/photonics10010007
Publisher site
See Article on Publisher Site

Abstract

hv photonics Article Chromatic-Aberration-Corrected Hyperspectral Single-Pixel Imaging 1 1 , 2 , 3 1 1 Ying Liu , Zhao-Hua Yang * , Yuan-Jin Yu * , Ling-An Wu , Ming-Yue Song and Zhi-Hao Zhao School of Instrumentation and Optoelectronics Engineering, Beihang University, Beijing 100191, China School of Automation, Beijing Institute of Technology, Beijing 100081, China Institute of Physics, Chinese Academy of Sciences, Beijing 100190, China * Correspondence: yangzh@buaa.edu.cn (Z.-H.Y.); yuanjin.yu@bit.edu.cn (Y.-J.Y.) Abstract: With the emerging development of hyperspectral single-pixel imaging (SPI) systems, the trade-off between the simplicity of optical structure and the correction of chromatic aberration is now an essential factor to be considered. To address both issues simultaneously, we propose a chromatic-aberration-corrected hyperspectral single-pixel imaging scheme, which is based on spectral measurement and dispersion correction. Its achromatism feature is evaluated by optical simulations and proof-of-concept experiments. Moreover, to overcome the shortcomings of traditional algorithms, a new adaptive iterative algorithm is employed, which can further optimize image quality. The results demonstrate that both dispersion and noise in our system are significantly reduced. Taking the position coordinate variance as a figure of merit, we have realized an order of magnitude improvement in the lateral chromatic aberration over the spectral range of 400–780 nm compared to that in conventional hyperspectral SPI. Meanwhile, the contrast-to-noise ratio in our system is enhanced on average by 3 dB. To the best of our knowledge, this is the first such demonstration, and the technique presents possibilities for future integrated applications of high spatial/spectral resolution over the entire visible range, and the system has the potential to be scaled down for future integrated applications. Keywords: single-pixel imaging; hyperspectral; chromatic aberration correction; optical system 1. Introduction Citation: Liu, Y.; Yang, Z.-H.; Yu, Y.-J.; The first recorded attempt at single-pixel imaging may be traced back to 1884, when Wu, L.-A.; Song, M.-Y.; Zhao, Z.-H. Paul Nipkow patented the flying-spot camera, using a Nipkow disk to modulate the Chromatic-Aberration-Corrected light source [1]. After years of development, contemporary single-pixel imaging (SPI) has Hyperspectral Single-Pixel Imaging. Photonics 2023, 10, 7. https:// advanced in parallel with computational ghost imaging (GI) [2,3]. The basic idea is to doi.org/10.3390/photonics10010007 modulate the reflected/transmitted beam illuminating an object with a sequence of patterns and record the corresponding total intensity values with a single-pixel (bucket) detector, Received: 28 October 2022 then recover the image through convolution. The patterns are generally pre-ordered and Revised: 18 December 2022 generated by a programmable spatial light modulator (SLM). The measured data, combined Accepted: 19 December 2022 with the prior information of the illumination speckles, constitute the essential elements of Published: 22 December 2022 a given computational reconstruction algorithm. Currently, there is a variety of reconstruction algorithms. Ferri et al. first proposed differential GI in 2010 [4], which has been widely adopted for its intrinsic advantage of anti- Copyright: © 2022 by the authors. noise interference. Later, many other algorithms were proposed, for instance, normalized Licensee MDPI, Basel, Switzerland. GI [5], high-order GI [6], correspondence GI [7] and self-evolving GI [8]. This article is an open access article In single-pixel imaging, only the total intensity values need to be measured, which distributed under the terms and makes it particularly suitable for applications in low visibility conditions, such as in remote conditions of the Creative Commons sensing [9,10] and underwater imaging [11]. Formerly, single-pixel sensing operated on Attribution (CC BY) license (https:// a single wavelength. When the process is extended to a wide spectral range, chromatic creativecommons.org/licenses/by/ aberration occurs since the same material exhibits different refractive indices at different 4.0/). Photonics 2023, 10, 7. https://doi.org/10.3390/photonics10010007 https://www.mdpi.com/journal/photonics Photonics 2023, 10, 7 2 of 12 wavelengths. This type of deviation would restrict the further enhancement of the spatial resolution of a broadband single-pixel imager. More severely, the larger the spectral range, the larger the chromatic aberration would be. Currently, known solutions, including temporal multiplexing [12] and frequency-division multiplexing [13], which are akin to introducing optical bandpass filters, predominantly suffer from the redundancy of optical structure and restricted spectral resolution. Schemes of coded aperture snapshot spectral imagers [14–16] could solve the problem, but with a significantly increased complexity of the algorithms. Hyperspectral imaging has been implemented based on SPI with a flat-field grating [17–19]. However, these schemes paid little attention to chromatic aberration since they were basically proximity range detection. To minimize the impact of chromatic aberration and to reduce structural vulnerability, spectral anti-aliasing with as simple as possible optical arrangements is required. In traditional spectral systems, this is achieved by dispersing the incident polychromatic light before the imaging section with an independent set of devices for each waveband, but these systems are always heavy and cumbersome. In this article, we propose a solution based on spectral measurement and dispersion correction, which we call chromatic-aberration-corrected hyperspectral single-pixel imaging (CHSPI). A spectrometer is used as the single-pixel sensor, so as to measure the spectrum of the reflected light. A closed-circuit television camera lens (later referred to as the TV lens) with a double Gaussian structure is adopted instead of a single thin lens. To eliminate the ripple noise when the object is imaged in a real light field and reconstructed by a traditional SPI algorithm, an adaptive and iterative correction algorithm is formulated. Simulations and a series of proof-of-concept experiments are conducted to test the achromatic capability of our scheme. As far as we are aware, this is the first demonstration of chromatic aberration correction in broadband hyperspectral SPI. The lateral chromatic aberration is suppressed by an order of magnitude over 400–780 nm in terms of pixel coordinate variance. Compared with previous schemes, our system can achieve high-contrast achromatic hyperspectral imaging over a broad range while the structure is lightweight and stable—characteristics that are particularly desirable in remote sensing applications. 2. System Model and Theoretical Derivation 2.1. Principle of Hyperspectral Differential Ghost Imaging In our scheme, a differential GI-based algorithm was selected to reconstruct the image while suppressing the noise [4]. In this linear non-iterative method, a variable approximate to the average intensity of the light field is subtracted from each pixel. In this way, the effective signal of the bucket detector just reflects the fluctuations of the light so that the noise is significantly suppressed and the signal-to-noise ratio (SNR) of the reconstructed Photonics 2022, 9, x FOR PEER REVIEW 3 of 14 image is improved. The basic principle of SPI is shown in Figure 1, where SLM is a spatial light modulator. Figure 1. Principle of single-pixel imaging. Figure 1. Principle of single-pixel imaging. In traditional SPI, the image is restored by correlating the light field modulation func- tion I(x) with the bucket signal S: (1) SI = ()xxR()dx where R(x) is the (intensity) reflection function of the object, and is the plane on which the image is modulated. The retrieved image is thus: G()xx = δδSI ⋅ () (2) 1 N δII ()xx=− () I(x) where denotes the ensemble average and . In differ- ... = (...) i =1 δRR ()xx=− () R ential GI, the fluctuating part of R(x) is defined as , where R=( I()xx Rxx)d I( )d x  is the average transmission function of the object. We de- fine a new differential bucket signal as: SI = ()xx δR()dx (3) Suppose NN × is the number of pixels in each pattern and K is the number of pre- set 2D spatial patterns (KN ≤×N ). Since the shot noise is negligible compared with the NN ××K abundant detected photons, the known modulation matrix M x ∈ and the beam () NN ××K I x ∈ , which is reflected by the SLM, are perfectly correlated variables; we can () IM ()xx = α ( ) introduce a factor to describe the correlation as . At the same time, we SM = ()xxd can set a virtual bucket signal to simplify the expression. Notice that can be expressed in terms of as: R = (4) α S In this way, S1 can be measured from the following operative form SS =− S (5) 1 R The image is then recovered by substituting S with S1 in Equation (2), namely, Photonics 2023, 10, 7 3 of 12 In traditional SPI, the image is restored by correlating the light field modulation function I(x) with the bucket signal S: S = I(x)R(x)dx (1) where R(x) is the (intensity) reflection function of the object, and A is the plane on which the image is modulated. The retrieved image is thus: G(x) = dS dI(x) (2) h i where . . . = (. . .) denotes the ensemble average and dI(x) = I(x) I(x) . h i å h i N i=1 In differential GI, the fluctuating part of R(x) is defined as dR(x) = R(x) R, where R R 2 2 R = hI(x)iR(x)d x/ hI(x)id x is the average transmission function of the object. We define a new differential bucket signal S as: S = I(x)dR(x)dx (3) Suppose N N is the number of pixels in each pattern and K is the number of preset 2D spatial patterns (K  N N). Since the shot noise is negligible compared with the NNK abundant detected photons, the known modulation matrix M(x) 2 R and the beam NNK I(x) 2 R , which is reflected by the SLM, are perfectly correlated variables; we can introduce a factor a to describe the correlation as I(x) = aM(x). At the same time, we can set a virtual bucket signal S = M(x)d x to simplify the expression. Notice that R can be expressed in terms of hS i as: 1 hSi R = (4) a S h i In this way, S can be measured from the following operative form h i S = S S (5) 1 R hS i The image is then recovered by substituting S with S in Equation (2), namely, G (x) = dS  dI(x) h i 1 1 (6) hSi = hS I(x)i hS I(x)i hS i In hyperspectral imaging, the above reconstruction of grayscale images corresponds to periodic sampling in the spectral dimension. More specifically, the detected intensity values over every 10 nm of the spectrum are summed as the “bucket detector value” S of the corresponding band I, from which the grayscale image G (x) is reconstructed as in 1i Equation (7) below: G (x) = hdS dI(x)i i 1i (7) hS i = hS  I(x)i hS  I(x)i i i hS i iR Based on the CIE 1931 Standard Colorimetric Observer, within the visible spectrum of 400–780 nm, the weighted sum of the images in each band is used to obtain the mixed color tristimulus image X(x), Y(x) and Z(x): X(x) = k G (x) x(l)Dl (8) å 1i Y(x) = k G (x) y(l)Dl (9) å 1i Z(x) = k G (x) z(l)Dl (10) å 1i where x(l), y(l), and z(l) are the color matching functions, k is the adjustment coefficient, and Dl is a fixed wavelength interval [20]. Then, by respectively plugging the three Photonics 2022, 9, x FOR PEER REVIEW 4 of 14 G()xx = δδSI ⋅ () 1 1 (6) =⋅SII ()xx − S⋅ ( ) In hyperspectral imaging, the above reconstruction of grayscale images corresponds to periodic sampling in the spectral dimension. More specifically, the detected intensity values over every 10 nm of the spectrum are summed as the “bucket detector value” G () x 1i of the corresponding band I, from which the grayscale image is reconstructed as in Equation (7) below: G (x) = δδ S ⋅ I() x 1i 1i S (7) =⋅ SII ()xx − S⋅ () ii iR Based on the CIE 1931 Standard Colorimetric Observer, within the visible spectrum of 400‒780 nm, the weighted sum of the images in each band is used to obtain the mixed color tristimulus image X(x), Y(x) and Z(x): XG ()xk=⋅ ()xx(λλ)Δ (8)  1i YG ()xk=⋅ ()xy(λλ)Δ (9) 1i ZG ()xk=⋅ ()xz(λλ)Δ (10)  1i Photonics 2023, 10, 7 4 of 12 where x() λ , y() λ , and z() λ are the color matching functions, k is the adjustment coef- ficient, and Δλ is a fixed wavelength interval [20]. Then, by respectively plugging the three matrices X(x), Y(x), and Z(x) into the three channels of the resultant RGB image, the matrices X(x), Y(x), and Z(x) into the three channels of the resultant RGB image, the entire entire reconstruction process can be completed. reconstruction process can be completed. 2.2. Principle of Chromatic Aberration Correction 2.2. Principle of Chromatic Aberration Correction The refractive index of a given material varies with the wavelength of light. Therefore, The refractive index of a given material varies with the wavelength of light. There- in a typical single thin lens imaging system, the object is imaged at different locations fore, in a typical single thin lens imaging system, the object is imaged at different locations when illuminated by polychromatic light; in other words, the image would be aliased when illuminated by polychromatic light; in other words, the image would be aliased into into defocused speckles at any position [20], as shown in Figure 2. Thus, this chromatic defocused speckles at any position [20], as shown in Figure 2. Thus, this chromatic aber- aberration would decrease the accuracy of R(x) in Equation (1). ration would decrease the accuracy of R(x) in Equation (1). Figure 2. Schematic of chromatic aberration along the optical axis. A: object beam. Wavelengths of F, D, and C are 486 nm, 588 nm, and 656 nm, respectively. To mitigate the problem, achromatic design is required. For a visual system, D light (588 nm) is generally chosen as the reference wavelength. The imaging positions of F light (486 nm) and C light (656 nm) are designed to coincide with the reference. The deviation between the imaging positions of two colors on the optical axis is called positional chromatic aberration, represented by 2 i 0 0 0 DL = L L = (11) FC F C 2 2 n0 u0 0 0 where h is the projection height of the object on each lens, n and u are the refractive index and aperture angle on the image side, and j and n are the focal power and Abbe coefficient i i of each lens, respectively [20]. To eliminate the chromatic aberration, h is expected to be zero. It can be seen that a single lens cannot correct the chromatic dispersion because its focal power, j, is always a non-zero constant but may be negative or positive depending on whether the lens 2 i is convex or concave, respectively. By selecting a suitable combination of lenses, h can be reduced to zero. Our scheme uses broadband light to illuminate a linear array detector for detection, so eliminating chromatic aberration is a key issue. Meanwhile, considering the advantage of reducing vertical-axis aberration (such as distortion, etc.), the double Gaussian lens is used. The double Gaussian lens is a typical successful adoption of the above combination, as shown in Figure 3. Two meniscus lenses, one positive meniscus L and another negative meniscus L , can reduce the spherical aberration and longitudinal chromatic aberration by 2 Photonics 2022, 9, x FOR PEER REVIEW 5 of 14 Figure 2. Schematic of chromatic aberration along the optical axis. A: object beam. Wavelengths of F, D, and C are 486 nm, 588 nm, and 656 nm, respectively. To mitigate the problem, achromatic design is required. For a visual system, D light (588 nm) is generally chosen as the reference wavelength. The imaging positions of F light (486 nm) and C light (656 nm) are designed to coincide with the reference. The deviation between the imaging positions of two colors on the optical axis is called positional chro- matic aberration, represented by ν (11) Δ= L' L' − L' =− FC F C n' u' where h is the projection height of the object on each lens, n' and u' are the refractive index and aperture angle on the image side, and ϕ and ν are the focal power and i i Abbe coefficient of each lens, respectively [20]. 2 i To eliminate the chromatic aberration, h is expected to be zero. It can be seen that a single lens cannot correct the chromatic dispersion because its focal power, ϕ , is always a non-zero constant but may be negative or positive depending on whether the lens is convex or concave, respectively. By selecting a suitable combination of lenses, h can be reduced to zero. Our scheme uses broadband light to illuminate a linear array detector for detection, so eliminating chromatic aberration is a key issue. Meanwhile, considering the advantage of reducing vertical-axis aberration (such as distortion, etc.), the double Gaussian lens is used. The double Gaussian lens is a typical successful adoption of the above combination, Photonics 2023, 10, 7 5 of 12 as shown in Figure 3. Two meniscus lenses, one positive meniscus L1 and another negative meniscus L2, can reduce the spherical aberration and longitudinal chromatic aberration by splitting one positive lens into two so that the curvature is spread over four surfaces splitting one positive lens into two so that the curvature is spread over four surfaces instead instead of two, while the overall power of the lens is preserved. Subsequently, placing of two, while the overall power of the lens is preserved. Subsequently, placing lenses L lenses L3 and L4 to face L1 and L2 symmetrically would notably diminish the vertical axis and L to face L and L symmetrically would notably diminish the vertical axis aberration, 4 1 2 aberration, such as distortion. The vital step to correct the chromatic aberration is through such as distortion. The vital step to correct the chromatic aberration is through replacing replacing the negative meniscus lenses L2 and L3 with hyperchromatic lenses L21, L22 and the negative meniscus lenses L and L with hyperchromatic lenses L , L and L , L , 2 3 21 22 31 32 L31, L32, respectively. A hyperchromatic lens is a cemented doublet lens with the same respectively. A hyperchromatic lens is a cemented doublet lens with the same refractive refractive index at the given wavelength but different dispersion, such that changing the index at the given wavelength but different dispersion, such that changing the radius of the radius of the curvature of the cemented surface does not alter the focal length of the pri- curvature of the cemented surface does not alter the focal length of the primary wavelength mary wavelength but adjusts the chromatic aberration [21]. By arranging the above lenses but adjusts the chromatic aberration [21]. By arranging the above lenses in proper positions, in proper positions, the final double Gaussian lens can correct chromatic aberration. the final double Gaussian lens can correct chromatic aberration. Photonics 2022, 9, x FOR PEER REVIEW 6 of 14 Figure 3. Double Gaussian structure. Figure 3. Double Gaussian structure. Figure 3. Double Gaussian structure [21]. 3. Simulation 3. Simulation A series of simulations were performed to compare the capability to correct for chro- A series of simulations were performed to compare the capability to correct for chro- matic aberration in conventional hyperspectral SPI (below referred to as HSPI) and our matic aberration in conventional hyperspectral SPI (below referred to as HSPI) and our CHSPI systems. The focal length and aperture of the two groups were equally set to 50 mm CHSPI systems. The focal length and aperture of the two groups were equally set to 50 and 25 mm, respectively. Figure 4 shows the general layouts, the lateral chromatic aberra- mm and 25 mm, respectively. Figure 4 shows the general layouts, the lateral chromatic tion curves, and the chromatic focal shift curves of the two systems. The lateral chromatic aberration curves, and the chromatic focal shift curves of the two systems. The lateral aberration is generally measured by the furthest deviation from the vertical axis in the field chromatic aberration is generally measured by the furthest deviation from the vertical axis and the chromatic focal shift by the maximum difference between the focal positions in the in the field and the chromatic focal shift by the maximum difference between the focal given wavelength range. As the diagrams in Figure 4c,d illustrate, the maximal absolute positions in the given wavelength range. As the diagrams in Figure 4c,d illustrate, the value of lateral chromatic aberration in the CHSPI system is only 1/7 of that in the HSPI maximal absolute value of lateral chromatic aberration in the CHSPI system is only 1/7 of system over the wavelength range of 400–780 nm and the corresponding chromatic focal that in the HSPI system over the wavelength range of 400‒780 nm and the corresponding shift 1/15. chromatic focal shift 1/15. Figure 4. Simulation results of a single-lens system and TV lens system. Layout of imaging parts in Figure 4. Simulation results of a single-lens system and TV lens system. Layout of imaging parts (a) hyperspectral SPI (HSPI) and (b) chromatic-aberration-corrected hyperspectral single-pixel im- in (a) hyperspectral SPI (HSPI) and (b) chromatic-aberration-corrected hyperspectral single-pixel aging (CHSPI). (c) Lateral chromatic aberration curve: field of view vs. lateral chromatic aberration. imaging (CHSPI). (c) Lateral chromatic aberration curve: field of view vs. lateral chromatic aberration. (d) The chromatic focal shift curve: wavelength vs. chromatic focal shift in the range from 486 nm of F-light to 656 nm of C-light. (d) The chromatic focal shift curve: wavelength vs. chromatic focal shift in the range from 486 nm of F-light to 656 nm of C-light. 4. Experimental Details To verify the effectiveness of dispersion correction of our hyperspectral correlation imaging experimental system, we conducted an experiment consisting of a system based on the TV lens and a control group based on a single lens. The dispersion correction per- formance of the two systems was evaluated by comparing the relative deviations of the reconstructed images of different spectral bands at corresponding points in the recovered single-pixel images of the two groups under the same environmental conditions. The schematic diagram and actual view of our experimental setup are illustrated in Figure 5. A broadband beam of light from a supercontinuum laser source (LEUKOS ROCK 400, 400–2400 nm) was expanded to illuminate the object, a 3 × 3 × 3 Rubik’s cube of size 54 × 54 × 54 mm , which was then imaged by a TV lens onto a common type of SLM—a digital micromirror device (DMD) (ViALUX V-9501, 1920 × 1080) prepro- grammed with a series of Hadamard matrices. After modulation, the beam was focused into a spectrometer (Ocean Optics FLAME-S-UV-VIS, 340–1100 nm). We employed 4096 speckle patterns generated under the cake-cutting basis Hadamard basis algorithm [22] for both the control and TV lens measurements. Photonics 2023, 10, 7 6 of 12 4. Experimental Details To verify the effectiveness of dispersion correction of our hyperspectral correlation imaging experimental system, we conducted an experiment consisting of a system based on the TV lens and a control group based on a single lens. The dispersion correction performance of the two systems was evaluated by comparing the relative deviations of the reconstructed images of different spectral bands at corresponding points in the recovered single-pixel images of the two groups under the same environmental conditions. The schematic diagram and actual view of our experimental setup are illustrated in Figure 5. A broadband beam of light from a supercontinuum laser source (LEUKOS ROCK 400, 400–2400 nm) was expanded to illuminate the object, a 3  3  3 Rubik’s cube of size 54  54  54 mm , which was then imaged by a TV lens onto a common type of SLM—a digital micromirror device (DMD) (ViALUX V-9501, 1920 1080) preprogrammed with a series of Hadamard matrices. After modulation, the beam was focused into a spectrometer (Ocean Optics FLAME-S-UV-VIS, 340–1100 nm). We employed 4096 speckle Photonics 2022, 9, x FOR PEER REVIEW 7 of 14 patterns generated under the cake-cutting basis Hadamard basis algorithm [22] for both the control and TV lens measurements. Figure 5. Schematic (a) and photograph (b) of the CHSPI setup. Figure 5. Schematic (a) and photograph (b) of the CHSPI setup. 5. Algorithm Optimization 5. Algorithm Optimization In the process of reconstructing the grayscale image of each band, erroneous black In the process of reconstructing the grayscale image of each band, erroneous black points (zero points) are prone to appear at the upper-left edge, forming noise ripples, as points (zero points) are prone to appear at the upper-left edge, forming noise ripples, as shown in Figure 6a. For a picture with a homogeneous background and a contrasting shown in Figure 6a. For a picture with a homogeneous background and a contrasting ob- object, these segments should have been part of and not darker than the background. As ject, these segments should have been part of and not darker than the background. As Equation (2) shows, the essence of GI computation is matrix multiplication. Under the Equation (2) shows, the essence of GI computation is matrix multiplication. Under the full-sampling condition, the solution should be unique. Because the modulation matrix is a full-sampling condition, the solution should be unique. Because the modulation matrix is constant known parameter, only various unpredictable noises can explain this error. To a constant known parameter, only various unpredictable noises can explain this error. To minimize these errors while retaining the original information as much as possible, several minimize these errors while retaining the original information as much as possible, sev- methods are adopted. According to the ripple characteristics, the pixels adjacent to the eral methods are adopted. According to the ripple characteristics, the pixels adjacent to error point are also prone to error, which excludes the traditional median or mean filtering the error point are also prone to error, which excludes the traditional median or mean methods of correction. Nor can frequency domain filtering be used since the ripples do not filtering methods of correction. Nor can frequency domain filtering be used since the rip- appear in all areas of the image. Therefore, on the basis of differential GI but to overcome ples do not appear in all areas of the image. Therefore, on the basis of differential GI but its inadequacies, we designed a new adaptive iterative correction algorithm, as illustrated to overcome its inadequacies, we designed a new adaptive iterative correction algorithm, in Figure 6b. as illustrated in Figure 6b. In each iteration, the gray value of the error point is set as the global gray scale average to eliminate its impact on the image contrast before re-normalization. The error point i i always preferentially appears at the position of (1, 2 ) or (2 , 1)(i = 1, 2, 3 . . . . . . ). After i i1 i i1 each correction, its position moves from (1, 2 ) to (1, 2 + 1) or from (2 , 1) to (2 + 1, 1) but never beyond the upper-left quarter of the image. We define the edge region as two line segments from (1, 2) to (1, N/2) and from (2, 1) to (N/2, 1) to simplify the computation. Each time, we search for the first zero point from left to right and from top to bottom. The whole process is finished when the first zero point no longer appears in the edge region. The SPI images of different spectral bands before and after correction are shown in Figure 6c. To quantitatively evaluate the image quality, we use the contrast-to-noise ratio Photonics 2023, 10, 7 7 of 12 (CNR) instead of the conventional signal-to-noise ratio because we do not have an original object image for comparison. The CNR is defined as [23]: CNR =10 log (S /s ) (12) where S is the variance of the object area, and s is the variance of the background area. o b An average increase of more than 3 dB in Figure 6c clearly exemplifies the improvement in Photonics 2022, 9, x FOR PEER REVIEW 8 of 14 image quality by our algorithm. Figure 6. Proposed adaptive iterative correction algorithm. (a) The noise ripples (zero points) on the Figure 6. Proposed adaptive iterative correction algorithm. (a) The noise ripples (zero points) on the upper-left edge. (b) The adaptive iterative correction algorithm used to correct the erroneous black upper-left edge. (b) The adaptive iterative correction algorithm used to correct the erroneous black noise ripples. (c) Comparison of SPI images in different spectral bands before and after correction. noise ripples. (c) Comparison of SPI images in different spectral bands before and after correction. In each iteration, the gray value of the error point is set as the global gray scale aver- 6. Results and Discussion age to eliminate its impact on the image contrast before re-normalization. The error point i i always preferentially appears at the position of (1, 2 ) or (2 , 1)(i = 1, 2, 3……). After each The optimized results of each wavelength band of hyperspectral SPI in the control i i−1 i i−1 correction, its position moves from (1, 2 ) to (1, 2 + 1) or from (2 , 1) to (2 + 1, 1) but never and experimental groups are shown in Figure 7, where pseudo-coloring has been applied. beyond the upper-left quarter of the image. We define the edge region as two line seg- In Figure 8, under the CIE1931 standard observer model, images of all spectral bands are ments from (1, 2) to (1, N/2) and from (2, 1) to (N/2, 1) to simplify the computation. Each fused to obtain the reconstructed color images. In the control group of Figure 8a, we can time, we search for the first zero point from left to right and from top to bottom. The whole see the influence of lateral chromatic aberration, as images are only clear from 580–610 nm, process is finished when the first zero point no longer appears in the edge region. The SPI elsewhere being blurred to varying degrees. By contrast, in the aberration-corrected images images of different spectral bands before and after correction are shown in Figure 6c. To in Figure 8b, the block margins are sharply outlined and distinct from the background quantitatively evaluate the image quality, we use the contrast-to-noise ratio (CNR) instead for all the bands; in addition, we can even discern the patchy reflections from the uneven of the conventional signal-to-noise ratio because we do not have an original object image surfaces of the colored squares, which are actually composed of glossy paper glued onto a for comparison. The CNR is defined as [23]: black cube. CNR=10log () S / σ (12) 10 ob where S is the variance of the object area, and σ is the variance of the background o b area. An average increase of more than 3 dB in Figure 6c clearly exemplifies the improve- ment in image quality by our algorithm. Photonics 2023, 10, 7 8 of 12 Photonics 2022, 9, x FOR PEER REVIEW 10 of 14 Figure 7. Reconstructed hyperspectral images over 400‒780 nm in the control (a) and TV lens (b) Figure 7. Reconstructed hyperspectral images over 400–780 nm in the control (a) and TV lens (b) groups, with their composite-colored images in the last figure. (c,d): Enlargements of the respective groups, with their composite-colored images in the last figure. (c,d): Enlargements of the respective images in the red boxes of (a,b) reveal that sharp edges are only seen in the yellow bands of the images single-lens in the syst red em, while boxes they of (a ,are di b) reveal stinctthat oversharp the vast edges majority of band are only seen range of the TV in the yellow system. bands of the single-lens system, while they are distinct over the vast majority of band range of the TV system. Photonics 2023, 10, 7 9 of 12 Photonics 2022, 9, x FOR PEER REVIEW 11 of 14 Figure 8. Reconstructed colored images of the control group (a) and the TV lens group (b). The white Figure 8. Reconstructed colored images of the control group (a) and the TV lens group (b). The white boxes show the main differences: purple or yellow haloes appear at the edges of the color blocks in boxes show the main differences: purple or yellow haloes appear at the edges of the color blocks in (a), while in (b), even details such as the white edges in boxes B2 and C2 are clearly revealed. (a), while in (b), even details such as the white edges in boxes B2 and C2 are clearly revealed. In order to quantify the performance of the two schemes, we examine a cross-section profile in the two resultant images (shown by the red lines in Figure 9a,c). The pixels’ gray values on the red line in each spectral band of 400–780 nm are successively extracted from left to right and plotted in the form of line charts in Figure 9b,d. Among them, six spectral bands (440 nm for blue, 490 nm for cyan, 530 nm for green, 580 nm for yellow, 630 nm for orange, and 680 nm for red) are selected. In the two sets of figures, three groups of pixels 1 , 2 , and 3 at the object’s edge positions are chosen where there are marked discontinuous changes in the gray values. Group corresponds to the edge of the white block and the maximum gray value point in the region, while group 2 in the center of the black gap corresponds to the minimum gray value. Group in the TV lens group of Figure 9c corresponds to the reflected white edge and the minimum gray value in the region, but in the control group of Figure 9a, this detail is too blurred, so the edge of the orange block is chosen, which corresponds to the maximum gray value in this region. In Figure 9b,d, each group is labelled and connected across the spectral bands with a blue line, while the vertical red dotted lines denote the abscissa of the chosen points with the first band (440 nm) as the reference origin. As the wavelength increases in Figure 9b, the corresponding points in the same group move to the left due to lateral chromatism. Additionally, the distance between groups 1 and 2 is significantly different in the change from blue to red; that is, from 3 pixels apart in the blue band to only 1 pixel apart in the red band, which also indicates the existence of lateral chromatism. However, in Figure 9d, we can see that the blue lines connecting the groups are straighter and almost coincide with the red lines, showing that aberration is corrected and variation with wavelength diminished. The overall deviations of the three positions from the reference origin point, as a function of wavelength, are shown in Figure 10a,b. In order to quantify the chromatic aberration, we calculated the pixel column number variances of the deviation at the three positions, as shown in Table 1. It can be seen that the Figure 9. Comparison of the performance of chromatic aberration correction. The intensity profiles variances in the TV lens scheme are on average more than an order of magnitude smaller across the red lines in the reconstructed images of (a) the control group and (c) the TV lens group than those in the single-lens scheme. vs. pixel number for different wave bands are shown in (b,d). Three groups of pixels ①, ②, and ③ are selected and in (b,d) are connected across the spectral bands with a blue line, while the vertical red dotted lines denote the abscissa of the chosen points with the first band (440 nm) as the reference origin. Photonics 2022, 9, x FOR PEER REVIEW 11 of 14 Figure 8. Reconstructed colored images of the control group (a) and the TV lens group (b). The white Photonics 2023, 10, 7 10 of 12 boxes show the main differences: purple or yellow haloes appear at the edges of the color blocks in (a), while in (b), even details such as the white edges in boxes B2 and C2 are clearly revealed. Figure 9. Comparison of the performance of chromatic aberration correction. The intensity profiles Figure 9. Comparison of the performance of chromatic aberration correction. The intensity profiles acr across the red lines in the reco oss the red lines in the reconstr nstructed images of ( ucted images of (a a)) the co the contr ntrol group and ( ol group and (c c) the TV le ) the TV lens ns group group vs. pixel number for different wave bands are shown in (b,d). Three groups of pixels ①, ②, and ③ vs. pixel number for different wave bands are shown in (b,d). Three groups of pixels 1 , 2 , and are selected and in (b,d) are connected across the spectral bands with a blue line, while the vertical are selected and in (b,d) are connected across the spectral bands with a blue line, while the Photonics 2022, 9, x FOR PEER REVIEW 12 of 14 red dotted lines denote the abscissa of the chosen points with the first band (440 nm) as the reference vertical red dotted lines denote the abscissa of the chosen points with the first band (440 nm) as the origin. reference origin. Figure 10. Pixel column number of the ①, ②, and ③ positions vs. wavelength over 400–780 nm for Figure 10. Pixel column number of the 1 , 2 , and 3 positions vs. wavelength over 400–780 nm for the control group (a) and the TV lens group (b). A flat line indicates no deviation in the focal point the control group (a) and the TV lens group (b). A flat line indicates no deviation in the focal point position, i.e. no lateral chromatic aberration occurs. position, i.e. no lateral chromatic aberration occurs. In order to quantify the chromatic aberration, we calculated the pixel column number variances of the deviation at the three positions, as shown in Table 1. It can be seen that the variances in the TV lens scheme are on average more than an order of magnitude smaller than those in the single-lens scheme. Table 1. Position Coordinate Variance in the Two Schemes. Scheme Name ① ② ③ HSPI scheme 4.8889 7.1137 1.8659 CHSPI scheme 0.1473 0.2459 0.2025 Improvement (HSPI/CHSPI) 33.1901 28.9292 9.2143 The theoretical spatial resolution limits of conventional imaging and the achieved resolutions of the two SPI systems are compared below to quantify their performance from another perspective. The apertures DC and DT of the single lens and TV lens are, respectively, 30 mm and 16 mm. The smallest unit that was clearly resolved by the single lens system was the 4 mm width of the black gaps between the colored blocks in A1 of Figure 8a, whereas the TV lens system could clearly resolve the white reflective strips in B2 and C2 of Figure 8b, which have a width of less than 1 mm. The distance between the object and the lens is more than 45 cm. Thus, the theoretical and actual angular resolutions can be derived as follows. For a central wavelength λ of 550 nm the theoretical resolution limit of conventional imaging using a single-lens system is ϕλ =1.22 / D =× 1.22 550nm/30mm=0.022mrad (13) CC while that of the TV lens system is ϕλ =1.22 / D =× 1.22 550nm/16mm=0.042mrad (14) TT The spatial resolution actually achieved in the single-lens SPI setup was (15) ϕ ≈ΔlL / = 4mm/450mm =8.9mrad while that in the improved SPI system was (16) ϕ ≈ΔlL / =1mm/450mm = 2.2mrad Photonics 2023, 10, 7 11 of 12 Table 1. Position Coordinate Variance in the Two Schemes. 1 2 3 Scheme Name HSPI scheme 4.8889 7.1137 1.8659 CHSPI scheme 0.1473 0.2459 0.2025 Improvement (HSPI/CHSPI) 33.1901 28.9292 9.2143 The theoretical spatial resolution limits of conventional imaging and the achieved resolutions of the two SPI systems are compared below to quantify their performance from another perspective. The apertures D and D of the single lens and TV lens are, C T respectively, 30 mm and 16 mm. The smallest unit that was clearly resolved by the single lens system was the 4 mm width of the black gaps between the colored blocks in A1 of Figure 8a, whereas the TV lens system could clearly resolve the white reflective strips in B2 and C2 of Figure 8b, which have a width of less than 1 mm. The distance between the object and the lens is more than 45 cm. Thus, the theoretical and actual angular resolutions can be derived as follows. For a central wavelength l of 550 nm the theoretical resolution limit of conventional imaging using a single-lens system is j = 1.22l/D = 1.22 550 nm/30 mm = 0.022 mrad (13) C C while that of the TV lens system is j = 1.22l/D = 1.22 550 nm/16 mm = 0.042 mrad (14) T T The spatial resolution actually achieved in the single-lens SPI setup was j  Dl/L = 4 mm/450 mm = 8.9 mrad (15) while that in the improved SPI system was j  Dl/L = 1 mm/450 mm = 2.2 mrad (16) Above are merely the results of our recent proof-of-principle experiments at present, and the actual spatial resolution limits need to be explored in subsequent research. Al- though the current measured resolutions are far from the theoretical values, comparing Equations (15) and (16), we can see that the achieved spatial resolution of the optimized TV lens system is more than a factor of four better than that of the control setup, even though its aperture is much smaller. 7. Conclusions As the first reported demonstration of chromatic aberration correction in hyperspectral SPI, our scheme succeeded in achieving an order of magnitude improvement in image quality across a wide spectral range of 400–780 nm. The images are reconstructed via an im- proved adaptive and iterative correction single-pixel algorithm, which allows information across all bandwidths to be retrieved within a short period of time with an image quality enhancement of more than 3 dB on average. Our scheme could provide a technical solution for spectral de-aliasing in SPI settings for microscopy and general analysis while providing a compact, lightweight, stable solution for remote sensing and satellite applications. Author Contributions: Conceptualization, Y.L.; methodology, Y.L.; software, Y.L.; validation, Y.L., M.-Y.S., and Z.-H.Z.; investigation, Y.L.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, Z.-H.Y., Y.-J.Y., and L.-A.W.; visualization, Y.L.; funding acqui- sition, Z.-H.Y., Y.-J.Y., and L.-A.W. All authors have read and agreed to the published version of the manuscript. Photonics 2023, 10, 7 12 of 12 Funding: This work was supported in part by the National Natural Science Foundation of China under Grants 61973018, 62173039, and 61975229, the Civil Space Project under Grant D040301, the National Key Research and Development Program of China under Grant 2018YFB0504302, and the Defense Industrial Technology Development Program under grant JCKY2021602B036. Institutional Review Board Statement: Not applicable. Informed Consent Statement: Not applicable. Data Availability Statement: Not applicable. Conflicts of Interest: The authors declare no conflict of interest. References 1. Yin, S.; Lu, G.; Zhang, J.; Yu, F.; Mait, J.N. Kinoform-based nipkow disk for a confocal microscope. Appl. Opt. 1995, 34, 5695. [CrossRef] [PubMed] 2. Shapiro, J.H. Computational ghost imaging. Phys. Rev. A 2008, 78, 061802. [CrossRef] 3. Shapiro, J.H.; Erkmen, B.I.; Lvovsky, A. Ghost imaging: From quantum to classical to computational. Adv. Opt. Photonics 2010, 2, 405–450. 4. Ferri, F.; Magatti, D.; Lugiato, L.A.; Gatti, A. Differential ghost imaging. Phys. Rev. Lett. 2010, 104, 253603. [CrossRef] [PubMed] 5. Sun, B.Q.; Welsh, S.S.; Edgar, M.P.; Shapiro, J.H.; Padgett, M.J. Normalized ghost imaging. Opt. Express 2012, 20, 16892–16901. [CrossRef] 6. Chen, X.H.; Agafonov, I.N.; Luo, K.H.; Liu, Q.; Xian, R.; Chekhova, M.V.; Wu, L.-A. High-visibility, high-order lensless ghost imaging with thermal light. Opt. Lett. 2010, 35, 1166–1168. [CrossRef] 7. Luo, K.H.; Huang, B.Q.; Zheng, W.-M.; Wu, L.-A. Nonlocal Imaging by Conditional Averaging of Random Reference Measure- ments. Chin. Phys. Lett. 2012, 29, 074216. [CrossRef] 8. Liu, B.; Wang, F.; Chen, C.; Dong, F.; McGloin, D. Self-evolving ghost imaging. Optica 2021, 8, 1340–1349. [CrossRef] 9. Erkmen, B.I. Computational ghost imaging for remote sensing. J. Opt. Soc. Am. A Opt. Image Sci. Vis. 2012, 29, 782–789. [CrossRef] 10. Yang, H.; Huang, Y.-H.; Zhong, C.; Shao, X.-P. Impact of Satellite Movement on Compressive Coded Aperture Spectral Imaging. Spacecr. Recovery Remote Sens. 2013, 34, 16–24. 11. Le, M.; Gao, W.; Zheng, H.; Liu, J.; Zhuo, X. Underwater computational ghost imaging. Opt. Express 2017, 25, 22859. [CrossRef] 12. Wang, Y.; Suo, J.; Fan, J.; Dai, Q. Hyperspectral Computational Ghost Imaging via Temporal Multiplexing. IEEE Photonics Technol. Lett. 2016, 28, 288–291. [CrossRef] 13. Bian, L.; Suo, J.; Situ, G.; Li, Z.; Fan, J.; Chen, F.; Dai, Q. Multispectral imaging using a single bucket detector. Sci. Rep. 2016, 6, 24752. [CrossRef] [PubMed] 14. Wagadarikar, A.A.; John, R.; Willett, R.; Brady, D. Single disperser design for coded aperture snapshot spectral imaging. Appl. Optics 2008, 47, B44–B51. [CrossRef] [PubMed] 15. Arguello, H.; Arce, G. Code aperture design for compressive spectral imaging. In Proceedings of the 2010 European Signal Processing Conference (EUSIPCO-2010), Aalborg, Denmark, 23–27 August 2010. 16. Arguello, H.; Arce, G.R. Rank Minimization Code Aperture Design for Spectrally Selective Compressive Imaging. IEEE Trans. Image Process. 2013, 22, 941–954. [CrossRef] [PubMed] 17. Chao, W.; Liu, X.F.; Yu, W.K.; Yao, X.R.; Li, L.Z.; Zhao, Q.; Zhai, G.-J. Compressed spectral imaging with a spectrometer. Opt. Commun. 2015, 352, 45–48. 18. Liu, S.; Liu, Z.; Wu, J.; Li, E.; Hu, C.; Tong, Z.; Shen, X.; Han, S. Hyperspectral ghost imaging camera based on a flat-field grating. Opt. Express 2018, 26, 17705–17716. [CrossRef] [PubMed] 19. Liu, B.; Wang, F.; Chen, C.; McGloin, D. Single-Pixel Diffuser Camera. IEEE Photonics J. 2021, 13, 1–5. [CrossRef] 20. Yu, D.-Y.; Tan, H.-Y. Engineering Optics, 4th ed.; China Machina Press: Beijing, China, 2015; pp. 93–94, 126–129. 21. Mandler, W. Design of Basic Double Gauss Lenses. In Proceedings of the 1980 International Lens Design Conference, Oakland, America, 16 September 1980. 22. Yu, W.K. Super sub-Nyquist single-pixel imaging by means of cake-cutting Hadamard basis sort. Sensors 2019, 19, 4122. [CrossRef] [PubMed] 23. Li, M.-F.; Yan, L.; Yang, R.; Kou, J.; Liu, Y.-X. Turbulence-free intensity fluctuation self-correlation imaging with sunlight. Acta Phys. Sin. 2019, 68, 094204. [CrossRef] Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Journal

PhotonicsMultidisciplinary Digital Publishing Institute

Published: Dec 22, 2022

Keywords: single-pixel imaging; hyperspectral; chromatic aberration correction; optical system

There are no references for this article.