Open Access
13 April 2023 Single-shot photoacoustic imaging with single-element transducer through a spatiotemporal encoder
Author Affiliations +
Abstract

Significance

Current photoacoustic (PA) imaging modalities typically require either serial detection with a single-element transducer or parallel detections with an ultrasonic array, indicating a dilemma between system cost and imaging throughput. PA topography through ergodic relay (PATER) was recently developed to address this bottleneck. However, PATER requires object-specific calibration due to varied boundary condition and must be recalibrated through pointwise scanning for each object before measurements, which is time-consuming and severely limits practical application.

Aim

We aim to develop a new single-shot PA imaging technique that only requires a one-time calibration for imaging different objects using a single-element transducer.

Approach

We develop an imaging method, PA imaging through a spatiotemporal encoder (PAISE), to address the above issue. The spatial information is effectively coded into unique temporal features by the spatiotemporal encoder, which allows for compressive image reconstruction. An ultrasonic waveguide is proposed as a critical element to guide the PA waves from the object into the prism, which effectively accounts for the varied boundary condition of different objects. We further add irregular-shaped edges on the prism to introduce randomized internal reflections and further facilitate the scrambling of acoustic waves.

Results

The proposed technique is validated through comprehensive numerical simulations and experiments, and it is demonstrated that PAISE can successfully overcome the changed boundary condition and can image different samples given a single calibration.

Conclusions

The proposed PAISE technique is capable of single-shot widefield PA imaging with a single-element transducer and does not require sample-specific calibration, which successfully overcomes the major limitation of previous PATER technology.

1.

Introduction

Optical imaging reveals important molecular and structural information in biological tissues. Microscopy technologies, such as confocal microscopy, multiphoton microscopy, and optical coherence tomography, can form high-resolution images through millimeter depth by detecting backscattered or fluorescent photons.14 Diffuse optical imaging technologies, such as diffuse optical tomography and spatial frequency-domain imaging, can probe deeper in biological tissues at the cost of lower resolution due to the nature of photon diffusion.57 As a complementary tool, photoacoustic (PA) imaging has a unique advantage of penetrating centimeters in biological tissues while maintaining high spatial resolution. With sensitivity toward optical absorption, PA imaging can probe both molecular and structural information, which holds promise for a wide range of applications.811

To date, most PA imaging systems require either point-by-point scanning using a single-element ultrasonic transducer or parallel detections using an ultrasonic transducer array. While the former is typically time-consuming and has a limited imaging throughput, the latter is complex and expensive.1115

Here, we propose a low-cost high-throughput PA imaging technique enabled by a spatiotemporal encoder (SE) to image different objects with only a single-element ultrasonic transducer and a single laser shot and without the need of object-specific calibration. We refer to this technique as PA imaging through a spatiotemporal encoder (PAISE). Previously, Li et al. demonstrated the feasibility of PA imaging with a single-element transducer [i.e., PA topography through ergodic relay (PATER)]. However, PATER requires object-specific calibration for each imaging object. In other words, the system must be recalibrated through pointwise scanning for each object before the measurement, which is time-consuming and severely limits practical application.16 In contrast, PAISE is featured with object-independent calibration, and once calibrated, can be applied for single-shot imaging of different objects, which overcomes the major limitation of PATER technology.

2.

Photoacoustic Imaging Through a Spatiotemporal Encoder

PAISE differs from conventional PA imaging in terms of measurement mechanism and image reconstruction. On the one hand, PA microscopy (such as OR-PAM) focuses a short light pulse onto a small spot (e.g., a diffraction-limited point) of the object to be imaged. The illuminated spot absorbs light energy and converts it into heat and consequently emits PA wave due to thermoelastic expansion.17,18 Subsequently, a single-element focused transducer detects the PA wave. The object image is eventually acquired through pointwise scanning of the entire field-of-view (FOV), as shown in Fig. 1(a). On the other hand, as shown in Fig. 1(b), photoacoustic computed tomography (PACT) applies widefield illumination upon the imaging object and uses a transducer array to detect emitted PA waves from the entire FOV in parallel. The detected signals are then used to reconstruct the object image. In contrast, PAISE (as well as PATER) uses a single-shot widefield illumination and only a single-element transducer to capture the entire image without scanning. As shown in Fig. 1(c), the PA waves from the FOV are encoded by the SE. The encoded signal is recorded by the single-element transducer, and the image is recovered through compressive image reconstruction.

Fig. 1

Three formations of PA imaging. (a) Conventional PA microscopy utilizes a single-element transducer but requires pointwise scanning to form an image. (b) PACT applies widefield illumination and detects PA waves in parallel, but it requires ultrasound transducer array which is complex and expensive. (c) With an SE, the proposed PAISE is capable of single-shot widefield imaging using a single-element transducer without the need of object-specific calibration.

JBO_28_4_046004_f001.png

To achieve compressive image reconstruction, the information of different spatial locations needs to be projected onto a set of incoherent basis functions via spatiotemporal encoding. As shown in Fig. 2(a), given a single-element transducer and two objects at its symmetric locations, the transducer can detect PA waves from both objects, but is unable to distinguish the source of the signal. In PATER, a right-angle prism was proposed to break the symmetry. As shown in Fig. 2(b), the PA waves from the objects contain unique temporal features and can be distinguished from the combined signal through compressive image reconstruction. However, it requires pointwise calibration of the entire FOV for each new object, which is time-consuming and severely limits practical application. To address the bottleneck of object-specific calibration, we propose a spatiotemporal encoding strategy (i.e., PAISE), which features object-independent calibration. As shown in Fig. 2(c), PAISE uses an ultrasonic pipe combined with a right-angle prism to achieve unique spatiotemporal encoding of PA waves originating from different locations in the FOV. In addition, the ultrasonic pipe is placed at the corner of the right-angle prism, so the ultrasonic signals from the object can be sufficiently scrambled (i.e., encoded) in the time domain to facilitate image reconstruction.

Fig. 2

An SE allows for single-shot widefield imaging with single-element transducer. (a) A single-element transducer cannot distinguish objects at symmetric locations. (b) With a right-angle prism, previous PATER technique is capable of signal separation between two objects but requires object-specific calibrations. (c) The proposed PAISE is capable of unique signal separation between two objects through compressive image reconstruction and only requires a one-time calibration for different objects.

JBO_28_4_046004_f002.png

3.

Propagation of PA Waves in the Spatiotemporal Encoder

The key advantage of PAISE over PATER is that it does not require object-specific calibration. In other words, once calibrated, PAISE can be applied to different objects to achieve single-shot PA imaging with a single-element transducer. To illustrate the mechanism of such advantage, the propagation of PA waves in the proposed SE is simulated based on the k-wave toolbox,19 and representative ultrasound wave fields are shown in Fig. 3. In Fig. 3, the simulation was conducted in two-dimensional (2D) to give a side view of the propagation of PA waves. The ultrasonic pipe and the right-angle prism were both modeled as fused silica (with 5900  m/s speed of sound), and the surrounding medium was modeled as air (with 343  m/s speed of sound). To aid the visualization of the propagation of PA waves in the pipe and the prism, the lengths were set as 6 cm for the pipe and 3 cm for the prism. In addition, the voxel size was set as 0.1  mm3 for the simulation. As shown in Fig. 3(a), the object is placed on top of the ultrasonic pipe, and the single-element transducer is placed at the corner of the right-angle prism. At t=0  μs, the short light pulse illuminates the object and generates PA wave. The ultrasound wave fields from timepoint 0.5 to 1.5  μs clearly illustrate the propagation of the PA wave traveling from the object, through the ultrasound pipe, and into the right-angle prism. The recorded PA signal is also plotted, and the blue curves represent the arrived signal at different time points, while the entire waveform is shown in gray. The flat blue lines in Fig. 3(a) indicate that until t=1.5  μs, the PA wave has not yet arrived at the transducer. After arrival into the prism, the PA wave bounces within the prism and at the same time the transducer records a unique time series which encodes the spatial information of the source of the PA wave. It is noted that while the PA wave bounces inside the prism, part of the wave would find its way back into the ultrasonic pipe and travel toward the object, as shown in Fig. 3(b). When the ultrasound wave reaches the top of the pipe, it would interact with the boundary (including the original object) and then travel back toward the prism, as shown in Fig. 3(c). It is important to note that the ultrasound wave coming back from the top boundary is dependent on the boundary condition, and after entering the prism, would affect the temporal feature of the recorded signal. On the other hand, the signals record before the arrival of the reflected PA wave is independent from the boundary condition. Such mechanism is the key feature that enables object-independent calibration in PAISE. In contrast, note that in PATER, the recorded signal would be continuously affected by the boundary condition, which dictates recalibrations for each imaging object.

Fig. 3

Propagation of PA waves in the SE. (a) The generated PA wave travels along the ultrasonic pipe toward the right-angle prism. (b) The PA wave reverberates inside the prism, and part of the wave travels back into the ultrasonic pipe toward the object. (c) The reflected ultrasound wave, which is dependent on the boundary condition, travels back into the prism. The blue curves represent arrived signal at different time points, and the gray curves represent the entire waveform.

JBO_28_4_046004_f003.png

4.

Experimental Validations

4.1.

Comparison of Image Reconstruction Between PATER and PAISE

To further illustrate the capability of object-independent calibration, we compared the proposed PAISE against PATER for image reconstruction with numerical simulations.

Following the procedures described in Li et al.,16 the compressive image reconstruction was implemented with a two-step iterative shrinkage/thresholding (TwIST) algorithm.20 Specifically, the signal recorded on the single-element transducer, denoted as s, can be represented as a linear combination of the calibration basis, denoted as ki, as shown in Eq. (1):

Eq. (1)

s=i=1NkiPi,
where i, N, and Pi denote calibrated pixel index, the total number of calibrated pixel locations, and the pixel intensity of the widefield image, respectively. Equation (1) can be cast to a matrix form for simplicity: s=KP, where K=[k1,k2,,kN] and P=[P1,P2,,PN]. The image P can be recovered by minimizing the objective function in Eq. (2), where ΦTV(P) is the total variation regularization term, and λ is the regularization parameter:

Eq. (2)

P^=argminPsKP2+2λΦTV(P).

In Fig. 4, to compare the image reconstruction between PATER and PAISE, the simulation was conducted in 3D. In the simulation, the prism was modeled as an off-the-shelf right-angle prism. Different from 2D simulation, 3D simulation requires significantly more voxels and consequently challenges the RAM size of the PC as well as CPU runtime. As proof-of-concept, in the 3D simulation (with voxel size of 0.1  mm3), the lengths of the pipe and the prism were set as 30 and 5 mm, respectively. The transducer was set as 0.2 mm in size and placed in direct contact with the lower corner of the prism. In addition, we note that the size and position of the transducer can be moderately adjusted without affecting the calibration, but one can envision that if the transducer is as large as the prism, the calibration would not work very well, since the acoustic waves would not be able to sufficiently scramble inside the prism before being collected by the transducer.

Fig. 4

Comparison of PATER and PAISE for image reconstruction. (a) After pointwise calibration, PATER was not able to image new object due to changed boundary condition between the calibration and measurement objects. (b) After calibration, the proposed PAISE successfully recovered the image of the new object regardless of the changed boundary condition.

JBO_28_4_046004_f004.png

As shown in Fig. 4, first, the calibration object was placed on top of the right-angle prism and pointwise calibration was conducted. Then, a C-shaped object was placed on the prism for widefield measurement. As shown in Fig. 4(a), the PATER was not able to reconstruct the C-shaped object due to the changed boundary condition between the calibration and the measurement. In contrast, with the proposed PAISE, the C-shaped object was successfully recovered in image reconstruction, as shown in Fig. 4(b).

4.2.

Single-Shot Imaging of Different Objects with a Single Calibration

In addition to the numerical simulation, the proposed PAISE was further validated in experiment to demonstrate single-shot imaging using a single-element transducer with a single calibration.

A combination of customized right-angle prism (PS611, Thorlabs, Inc.; 2.5 cm right-angle edge length) and an ultrasonic pipe (#65-839, Edmund Optics; 1 cm hexagonal aperture, 17.5 cm length) was chosen to serve as the SE. Both the prism and the ultrasonic pipe were made of UV-fused silica, which had negligible attenuation over the acoustic detection pathlength. A custom-made unfocused pin transducer (30 MHz central frequency and 1 mm element size) was placed at the corner of the prism to collect ultrasound signal. To facilitate ultrasound transmission, ultrasound gel was applied between the object and the pipe, while a thin layer of polyester resin was added between the prism and the pin transducer. We note that although polyester resin would generally impede ultrasound transmission, counterintuitively, we found a thin layer of it to be very helpful in our experiment: since it does not dry over time, one would not need to add the coupling gel and position the transducer repeatedly. In addition, a nanosecond laser with 532 nm wavelength and 5 ns pulse width (INNOSAB IS8II-E, EdgeWave GmbH; 2 kHz repetition rate) was used as light source in the PAISE system. The laser beam was partially reflected by a beam splitter to a photodiode (DET36A, Thorlabs, Inc.) for the correction of fluctuations in pulse energy. In calibration, the laser beam was focused through a lens (LA1433-A, Thorlabs, Inc.; 2.54 cm diameter and 150 mm focal length) onto the calibration object for pointwise scanning. A motorized XY scanner (PLS-85, Physik Instrumente, Germany) was used for calibration in our experiment. Alternatively, a galvo-resonant scanner can also be utilized for improved scanning speed. In widefield measurement, the laser beam passed through an engineered diffuser (EDC-5-A-1r, RPC Photonics, Inc.; 5.5 deg divergence angle) to provide widefield illumination over the entire FOV. For data acquisition, a two-channel digitizer (ATS9350, AlazarTech, Inc.; 12 bits, 250  MS/s sampling rate, and 65,536 data points/acquisition sampling length) was used to record the PA signal and the photodiode output. The calibration and measurement objects were made of black rubber.

Figure 5(a) shows a picture of the key experimental setup, including the ultrasonic waveguide, right-angle prism, and pin transducer. While PATER used an off-the-shelf right-angle prism, in our experiment we added irregular-shaped edges on the right-angle prism (using a grinder) to introduce randomized internal reflections and further facilitate the scrambling of acoustic waves. We note that the customized right-angle prism was not used in simulation due to the challenges of accurate modeling of its irregular-shaped edges. In addition, to make such a customized prism, one simply needs to randomly grind (for example, use a bench grinder and carefully hold in hand when grinding to introduce randomness) groove/ridge patterns onto the original edges of the off-the-shelf right-angle prism, where the groove/ridge patterns should be considerably larger than the acoustic wavelength to allow reflection. We note that while both the off-the-shelf right-angle prism and the customized right-angle prism can achieve object-independent calibration (when combined with the ultrasonic waveguide), the customized prism has the benefits of improved resolution and image reconstruction. Pictures of the customized right-angle prism from different viewing-angles are shown in Fig. 5(b) to illustrate the irregular-shaped edges. In experiment, as shown in Fig. 5(c), the calibration object was placed on top of the ultrasonic pipe, and a focused laser beam was used to scan a 6×6  mm FOV on the object with 100-μm increments. For widefield measurement, as shown in Fig. 5(d), the calibration object was removed, and a C-shaped object was placed inside the calibrated FOV on top of the ultrasonic pipe. A broad laser beam was sent to illuminate the object and the PA signal was collected. Then, another widefield measurement was conducted for a T-shape object. As shown in Fig. 5(e), both objects were successfully recovered in the image reconstruction. The results demonstrated that once calibrated, the proposed PAISE was able to image different objects regardless of the changed boundary condition. In terms of time cost, the image reconstruction with TwIST algorithm took 3.6±0.1  s (repeated 10 times) using a desktop computer with Intel Xeon Silver 4210R CPU (2.40 GHz) and 128 GB RAM. In the future, the reconstruction process can potentially be significantly accelerated by developing a deep learning model that does not involve iterative optimization and can directly output the reconstructed image in a straightforward manner.2124

Fig. 5

Validation of PAISE with different measurement objects. (a) Picture of key experimental setup including the ultrasonic waveguide, right-angle prism, and pin transducer. (b) Pictures of the customized right-angle prism from different viewing-angles showing irregular-shaped edges. (c) The system was calibrated by pointwise scanning using a focused laser beam. The calibrated area was 6×6  mm and the scanning increments were 100  μm. (d) After calibration, a C-shaped object and a T-shaped object were placed on top of the ultrasonic pipe and measured with widefield laser beam. (e) Both objects were successfully recovered in the image reconstruction.

JBO_28_4_046004_f005.png

In addition, we note that the same experiments were also conducted for PATER. As expected, PATER was not able to reconstruct different objects after a single calibration due to changed boundary conditions (data not shown).

5.

Discussion

We have developed and validated PAISE as a new method for single-shot PA imaging with single-element transducer. This was enabled through the use of an SE, which was able to effectively transform spatial information into unique temporal features. The ultrasonic waveguide in combination with a right-angle prism allows for separation of varied boundary conditions from encoded PA signals. In comparison to the conventional PA technologies, we demonstrated that the proposed PAISE was able to translate encoded temporal signature into original spatial locations using a single transducer. We further compared the proposed PAISE against PATER through comprehensive simulations to demonstrate the advantage of object-independent calibration. Finally, we validated the PAISE in experiments and demonstrated single-shot imaging of different objects without the need of object-specific calibration.

PAISE enjoys a number of advantages over current PA modalities. Compared with popular PA microscopy techniques that also uses a single-element transducer, PAISE is able to perform single-shot imaging, while PA microscopy techniques require pointwise scanning over the entire FOV. PACT is able to conduct single-shot imaging and does not require pointwise scanning. However, PACT requires an ultrasonic transducer array and is known for being complex and expensive, whereas PAISE only requires a single-element transducer.11,14 In addition, while PATER can also conduct single-shot imaging using a single-element transducer, it requires recalibration for each measurement object due to varied boundary conditions. In contrast, PAISE only requires one-time calibration and can be applied to different measurement objects without recalibration. Furthermore, while PATER is limited for 2D imaging, it is important to note that PAISE can potentially achieve single-shot 3D imaging. This is because the depth information is intrinsically encoded in the time of arrival of the acoustic signal, and the 3D image of different objects can subsequently be recovered given calibration of the 3D volume.

Regarding the pipe selection, we would like to note that to address the limitation of PATER (i.e., repeated calibration for different objects), we took significant amount of time to explore various potential geometries of a new SE, such as rectangular ones, curved ones, triangular ones, etc. In the end, we found the presented version (i.e., a pipe on top of a right-angle prism) to be a very simple yet effective solution. In terms of parameter selection, the diameter of the pipe should be significantly smaller than the length of the right-angle prism, so the ultrasonic signals from the object can be effectively scrambled and encoded in the time domain. In addition, the length of the pipe should be much longer compared to the length of the right-angle prism. This is critical, since after the PA signal from the object arrives the right-angle prism, while it gets partially reflected back into the pipe for a round trip, the remaining PA signal, without being affected by the changed boundary condition, requires a sufficient time window to find its way to the transducer through multiple internal reflections. Essentially, the duration of that time window is determined by the length of the pipe (i.e., time for the round trip).

In terms of the shape of the utilized transducer, in principle, one can use both flat and curved transducers in the experiment. In the presented setup, the flat transducer can be placed in contact with the prism (through a thin layer of polyester resin) and efficiently record PA signals. In contrast, if using a curved transducer, such as a focused transducer, one would have to add more coupling materials between the transducer and the prism surface, which may affect the SNR of the recorded signals.

In terms of FOV, PATER demonstrated 8×6  mm2 FOV while the proposed PAISE demonstrated 6×6  mm2 FOV. The FOV of the proposed method can be extended using an ultrasonic pipe with a larger aperture. In addition, it is worth noting that the FOV of both methods can be extended by calibrating a larger area.

In terms of resolution, similar to PATER, the resolution limit of the proposed system is also fundamentally determined by the acoustic wavelength in the material of the encoder (e.g., fused silica) at the ultrasonic transducer’s central frequency. In addition to the in-depth analysis of imaging resolution included in the previous PATER work, we would like to add one more note regarding the transducer and the prism for optimizing resolution in experiment. To achieve optimized resolution, after entering the prism, the acoustic waves should be sufficiently scrambled, so the spatial information of the object can be effectively transformed (i.e., encoded) into unique temporal features of the collected time series. Specifically, on the one hand, the size of the transducer should be relatively small so the acoustic waves can be sufficiently scrambled in the SE before being collected by the transducer. On the other hand, if the size of the transducer is too small, only a tiny fraction of the acoustic waves can be collected, which would lead to deteriorated SNR and consequently decreased resolution. In our study, we found the 1 mm transducer to be feasible in experiment. In addition, the shape of the prism is also crucial for the resolution. In experiment, we found that when using an off-the-shelf right-angle prism, while achieving object-independent calibration, the system resolution was approximately the size of the transducer. To push the resolution toward the acoustic wavelength, in this study, we added irregular-shaped edges on the prism to introduce randomized internal reflections and further facilitate the scrambling of acoustic waves. The customized prism enabled resolution close to 0.2 mm (data not shown), which is significantly smaller than the size of the transducer.

This work has several important implications for scientific research and clinical applications. For example, brain cortex imaging is of fundamental significance for neuroscience.5,2528 Current methods such as two-photon microscopy and fluorescence imaging are typically limited by a small FOV and requires exogenous contrast agents. In contrast, the proposed PAISE can potentially provide both large FOV and label-free functional imaging. Furthermore, in terms of temporal resolution, given sufficient repetition rate of the light source, PAISE can potentially achieve kilohertz monitoring, which is similar to that of EEG methods,16 while the latter is known to have limited spatial resolution. Therefore, in the future, PAISE may be a powerful tool for brain science research. With regard to clinical applications, PAISE may also enable single-shot stain-free histopathology combined with UV wavelengths, which would be significantly faster than current UV-PAM modalities that requires pointwise scanning.15,29,30

In summary, we have developed PAISE, a new imaging modality that can conduct single-shot imaging using a single-element transducer without the need of object-specific calibration, which successfully overcomes the major limitation of previous PATER technology. In the future, PAISE can be potentially useful for a wide range of biomedical applications.

Disclosures

L.V.W. has financial interests in Microphotoacoustics, Inc., CalPACT, LLC, and Union Photoacoustic Technologies, Ltd., which did not support this work.

Acknowledgments

The authors gratefully thank Dr. Qifa Zhou for the customized pin transducer. The authors would also like to thank Dr. Konstantin Maslov, Dr. Junhui Shi, Dr. En Bo, Dr. Yang Li, Dr. Lei Li, and Dr. Shuai Na for helpful discussions. This work was sponsored by the National Institutes of Health Grants DP1 EB016986 (NIH Director’s Pioneer Award), R01 CA186567 (NIH Director’s Transformative Research Award), and R01 EB016963.

References

1. 

J. Jonkman et al., “Guidance for quantitative confocal microscopy,” Nat. Protoc., 15 1585 –1611 https://doi.org/10.1038/s41596-020-0313-9 1754-2189 (2020). Google Scholar

2. 

N. M. Israelsen et al., “Real-time high-resolution mid-infrared optical coherence tomography,” Light Sci. Appl., 8 (1), 11 https://doi.org/10.1038/s41377-019-0122-5 (2019). Google Scholar

3. 

Z. Qin et al., “Adaptive optics two-photon microscopy enables near-diffraction-limited and functional retinal imaging in vivo,” Light Sci. Appl., 9 (1), 79 https://doi.org/10.1038/s41377-020-0317-9 (2020). Google Scholar

4. 

L. Huang et al., “Recurrent neural network-based volumetric fluorescence microscopy,” Light Sci. Appl., 10 (1), 62 https://doi.org/10.1038/s41377-021-00506-9 (2021). Google Scholar

5. 

A. T. Eggebrecht et al., “Mapping distributed brain function and networks with diffuse optical tomography,” Nat. Photonics, 8 (6), 448 –454 https://doi.org/10.1038/nphoton.2014.107 (2014). Google Scholar

6. 

Y. Zhao et al., “Shortwave-infrared meso-patterned imaging enables label-free mapping of tissue water and lipid content,” Nat. Commun., 11 (1), 5355 https://doi.org/10.1038/s41467-020-19128-7 NCAOBW 2041-1723 (2020). Google Scholar

7. 

Y. Zhao et al., “Halftone spatial frequency domain imaging enables kilohertz high-speed label-free non-contact quantitative mapping of optical properties for strongly turbid media,” Light Sci. Appl., 10 (1), 245 https://doi.org/10.1038/s41377-021-00681-9 (2021). Google Scholar

8. 

L. V Wang and S. Hu, “Photoacoustic tomography: in vivo imaging from organelles to organs,” Science, 335 1458 –1462 https://doi.org/10.1126/science.1216210 SCIEAS 0036-8075 (2012). Google Scholar

9. 

L. V. Wang and J. Yao, “A practical guide to photoacoustic tomography in the life sciences,” Nat. Methods, 13 (8), 627 –638 https://doi.org/10.1038/nmeth.3925 1548-7091 (2016). Google Scholar

10. 

P. Zhang et al., “In vivo superresolution photoacoustic computed tomography by localization of single dyed droplets,” Light Sci. Appl., 8 (1), 36 https://doi.org/10.1038/s41377-019-0147-9 (2019). Google Scholar

11. 

S. Na et al., “Massively parallel functional photoacoustic computed tomography of the human brain,” Nat. Biomed. Eng., 6 584 –592 https://doi.org/10.1038/s41551-021-00735-8 (2021). Google Scholar

12. 

X. L. Deán-Ben and D. Razansky, “Adding fifth dimension to optoacoustic imaging: volumetric time-resolved spectrally enriched tomography,” Light Sci. Appl., 3 e137 https://doi.org/10.1038/lsa.2014.18 (2014). Google Scholar

13. 

J. Yao et al., “Multiscale photoacoustic tomography using reversibly switchable bacterial phytochrome as a near-infrared photochromic probe,” Nat. Methods, 13 (1), 67 –73 https://doi.org/10.1038/nmeth.3656 1548-7091 (2015). Google Scholar

14. 

L. Li et al., “Single-impulse panoramic photoacoustic computed tomography of small-animal whole-body dynamics at high spatiotemporal resolution,” Nat. Biomed. Eng., 1 (5), 0071 https://doi.org/10.1038/s41551-017-0071 (2017). Google Scholar

15. 

T. T. W. Wong et al., “Label-free automated three-dimensional imaging of whole organs by microtomy-assisted photoacoustic microscopy,” Nat. Commun., 8 (1), 1386 https://doi.org/10.1038/s41467-017-01649-3 NCAOBW 2041-1723 (2017). Google Scholar

16. 

Y. Li et al., “Snapshot photoacoustic topography through an ergodic relay for high-throughput imaging of optical absorption,” Nat. Photonics, 14 (3), 164 –170 https://doi.org/10.1038/s41566-019-0576-2 NPAHBY 1749-4885 (2020). Google Scholar

17. 

J. Xia, J. Yao and L. V. Wang, “Photoacoustic tomography: principles and advances,” Prog. Electromagn. Res., 147 1 –22 https://doi.org/10.2528/PIER14032303 PELREX 1043-626X (2014). Google Scholar

18. 

P. Beard, “Biomedical photoacoustic imaging,” Interface Focus, 1 (4), 602 –631 https://doi.org/10.1098/rsfs.2011.0028 (2011). Google Scholar

19. 

B. E. Treeby and B. T. Cox, “k-Wave: MATLAB toolbox for the simulation and reconstruction of photoacoustic wave fields,” J. Biomed. Opt., 15 (2), 021314 https://doi.org/10.1117/1.3360308 JBOPFO 1083-3668 (2010). Google Scholar

20. 

J. M. Bioucas-Dias and M. A. T. Figueiredo, “A new TwIST: two-step iterative shrinkage/thresholding algorithms for image restoration,” IEEE Trans. Image Process., 16 (12), 2992 –3004 https://doi.org/10.1109/TIP.2007.909319 IIPRE4 1057-7149 (2007). Google Scholar

21. 

H. Lan et al., “Deep learning enabled real-time photoacoustic tomography system via single data acquisition channel,” Photoacoustics, 22 100270 https://doi.org/10.1016/j.pacs.2021.100270 (2021). Google Scholar

22. 

Y. Zhao et al., “Deep learning model for ultrafast multifrequency optical property extractions for spatial frequency domain imaging,” Opt. Lett., 43 (22), 5669 https://doi.org/10.1364/OL.43.005669 OPLEDP 0146-9592 (2018). Google Scholar

23. 

Y. Zhao et al., “Direct mapping from diffuse reflectance to chromophore concentrations in multi-FX spatial frequency domain imaging (SFDI) with a deep residual network (DRN),” Biomed. Opt. Express, 12 (1), 433 –443 https://doi.org/10.1364/BOE.409654 BOEICL 2156-7085 (2021). Google Scholar

24. 

Y. Zhao et al., “Quantitative real-time pulse oximetry with ultrafast frequency-domain diffuse optics and deep neural network processing,” Biomed. Opt. Express, 9 (12), 5997 https://doi.org/10.1364/BOE.9.005997 BOEICL 2156-7085 (2018). Google Scholar

25. 

J. Yao et al., “High-speed label-free functional photoacoustic microscopy of mouse brain in action,” Nat. Methods, 12 (5), 407 https://doi.org/10.1038/nmeth.3336 1548-7091 (2015). Google Scholar

26. 

M. T. Colonnese et al., “Development of hemodynamic responses and functional connectivity in rat somatosensory cortex,” Nat. Neurosci., 11 (1), 72 –79 https://doi.org/10.1038/nn2017 NANEFN 1097-6256 (2008). Google Scholar

27. 

A. T. Winder et al., “Weak correlations between hemodynamic signals and ongoing neural activity during the resting state,” Nat. Neurosci., 20 (12), 1761 –1769 https://doi.org/10.1038/s41593-017-0007-y NANEFN 1097-6256 (2017). Google Scholar

28. 

Y. Hoshi, N. Kobayashi and M. Tamura, “Interpretation of near-infrared spectroscopy signals: a study with a newly developed perfused rat brain model,” J. Appl. Physiol., 90 (5), 1657 –1662 https://doi.org/10.1152/jappl.2001.90.5.1657 (2001). Google Scholar

29. 

T. T. W. Wong et al., “Fast label-free multilayered histology-like imaging of human breast cancer by photoacoustic microscopy,” Sci. Adv., 3 e1602168 https://doi.org/10.1126/sciadv.1602168 STAMCV 1468-6996 (2017). Google Scholar

30. 

R. Cao et al., “Label-free intraoperative histology of bone tissue via deep-learning-assisted ultraviolet photoacoustic microscopy,” Nat. Biomed. Eng., 7 124 –134 https://doi.org/10.1038/s41551-022-00940-z (2022). Google Scholar

Biography

Yanyu Zhao is currently an associate professor at Beihang University. He received his PhD from the Department of Biomedical Engineering at Boston University. He was a postdoctoral research fellow at Boston University and California Institute of Technology. His research interests focus on the development of novel biomedical imaging technologies. He is a member of SPIE and OSA.

Lihong V. Wang is the Bren Professor of Medical Engineering and Electrical Engineering at the Andrew and Peggy Cherng Department of Medical Engineering at California Institute of Technology and was formerly the Gene K. Beare Distinguished Professorship of Biomedical Engineering at Washington University in St. Louis. He is a Fellow of AAAS, AIMBE, Electromagnetics Academy, IAMBE, IEEE, NAI, OSA, and SPIE. An honorary doctorate was conferred on him by Lund University, Sweden. He was inducted into the National Academy of Engineering.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Yanyu Zhao and Lihong V. Wang "Single-shot photoacoustic imaging with single-element transducer through a spatiotemporal encoder," Journal of Biomedical Optics 28(4), 046004 (13 April 2023). https://doi.org/10.1117/1.JBO.28.4.046004
Received: 3 February 2023; Accepted: 21 March 2023; Published: 13 April 2023
Lens.org Logo
CITATIONS
Cited by 4 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Prisms

Calibration

Ultrasound transducers

Transducers

Image restoration

Photoacoustic imaging

Pipes

Back to Top