Open Access
8 May 2023 Field programmable gate array compression for large array multispeckle diffuse correlation spectroscopy
Author Affiliations +
Abstract

Significance

Diffuse correlation spectroscopy (DCS) is an indispensable tool for quantifying cerebral blood flow noninvasively by measuring the autocorrelation function (ACF) of the diffused light. Recently, a multispeckle DCS approach was proposed to scale up the sensitivity with the number of independent speckle measurements, leveraging the rapid development of single-photon avalanche diode (SPAD) cameras. However, the extremely high data rate from advanced SPAD cameras is beyond the data transfer rate commonly available and requires specialized high-performance computation to calculate large number of autocorrelators (ACs) for real-time measurements.

Aim

We aim to demonstrate a data compression scheme in the readout field-programmable gate array (FPGA) of a large-pixel-count SPAD camera. On-FPGA, data compression should democratize SPAD cameras and streamline system integration for multispeckle DCS.

Approach

We present a 192 × 128 SPAD array with 128 linear ACs embedded on an FPGA to calculate 12,288 ACFs in real time.

Results

We achieved a signal-to-noise ratio (SNR) gain of 110 over a single-pixel DCS system and more than threefold increase in SNR with respect to the state-of-the-art multispeckle DCS.

Conclusions

The FPGA-embedded autocorrelation algorithm offers a scalable data compression method to large SPAD array, which can improve the sensitivity and usability of multispeckle DCS instruments.

1.

Introduction

Monitoring of cerebral blood flow is key to understanding brain health and neuronal activity. Diffuse correlation spectroscopy (DCS) is an optical technique capable of measuring blood flow noninvasively using coherent near-infrared light.16 Coherent light introduced from the source to the scalp through deep tissue returns a speckle pattern at the detector, which fluctuates rapidly in intensity due to movement of the tissue, including the red blood cells therein. Measurement of the temporal intensity autocorrelation function (ACF), i.e., g2(τ), of the speckle yields a correlation time constant (τc), which can be used to quantify the blood flow index.

By increasing the source–detector separation over the scalp, DCS can capture photons scattered from deep into the cerebral cortex. However, the increase in separation sacrifices the signal-to-noise ratio (SNR) of the return signal, which is crucial for maximizing the precision of the g2 measurement. To increase the SNR of the g2(τ) curve [defined as the ratio between mean and standard deviation (STD) of g2(τ) calculated over several integration periods], a multispeckle DCS approach was proposed and demonstrated.69 In multispeckle DCS, each independent speckle field scattered from the tissue is coupled to an independent single-photon avalanche diode (SPAD) or to each pixel of the SPAD array, showing an SNR gain that scales with the square root of the number of pixels in the SPAD camera. Despite the scalability advantage offered by SPAD camera technology, the multispeckle DCS systems reported so far were limited by the small number of pixels in the SPAD array and the lack of camera-embedded processing. The latter has required external large computing resources to sustain the high frame rate of the camera and impeded scalability to higher resolutions and multiple camera systems fitted on the scalp.

State–of-the-art hardware autocorrelators (ACs) can only operate on a low number of channels, usually in the order of 10 to 100 channels, which does not allow coupling to high resolution of state-of-the-art SPAD cameras for real-time operation. The work in Ref. 10 was the first field-programmable gate array (FPGA) implementation of AC able to scale to 1032 channels interfacing to a 32×32 SPAD array, but the normalization of the per-pixel ACF was kept in software on the host PC. To achieve the SNR scaling with number of pixels, the per-pixel ACFs need to be averaged together into a single ACF also known as ensemble ACF. Since the ensemble ACF can only be performed after the normalization of each per-pixel ACF, the AC presented in Ref. 10 must stream the ACF and the raw aggregated photon count of each pixel to the host PC. This imposes a large data rate requirement, which is not compatible with the trend of increasing pixel resolution of SPAD sensors and therefore limits the potential for achieving higher SNR in real-time multispeckle DCS. To fully exploit the large number of channels in SPAD cameras, ACs can benefit from keeping the entire elaboration pipeline locally on the FPGA, including normalization and ensemble averaging and use the limited bandwidth to the host PC to stream out the SNR-scaled ensemble ACF.

Here we present a large array SPAD camera with 128 16-lag linear ACs embedded on an FPGA able to calculate 12,288 channels and output both per-pixel ACFs and compute the ensemble ACF from 192×64  pixels in real time. This represents a 12 times increase in number of channels than in prior works.6,7,10 We achieved an SNR gain of 110 over a single-pixel DCS system, which is more than threefold increase in the SNR gain shown in prior works.6,7 Although this work focuses on the FPGA-embedded processing for multispeckle DCS, our SPAD camera can operate in the multispeckle time-domain DCS with an unprecedented precision of 33 ps time-of-flight (TOF) resolution. This is beyond the precision offered by a gated approach that typically has nanosecond scale timestamp.

2.

Materials and Methods

We used a time-resolved complementary-metal-oxide semiconductor (CMOS) SPAD sensor composed of 192×128  pixels with a 33 ps TOF resolution and a 135-ns time-to-digital converter (TDC) in each pixel. The SPAD array was packaged into a camera module, referred to as QuantICAM.11 Part of the TDC is an 8-bit ripple counter which was repurposed to enable single-photon counting of up to 255 photons per frame exposure in each pixel. We used an Opal Kelly 7310 based on a Xilinx Artix-7 FPGA to control the sensor and to read out the chip through 64 data output I/Os. Figure 1 shows a photo of the QuantICAM SPAD array used in this work. Detailed system architecture of the QuantICAM SPAD array can be found in Ref. 11.

Fig. 1

QuantICAM SPAD array circuit board used in this work.

JBO_28_5_057001_f001.png

The sensor was read out in a rolling row readout scheme. Pairs of rows were read out in parallel from the center towards the top and bottom of the sensor. Each row was read out by 32 I/Os, each I/Os serially reading out data from 4 pixels. The maximum raw data rate from QuantICAM is 64×100  Mbit/s or 6.4  Gbit/s (800  MB/s), which is twice the data throughput of standard Universal Serial Bus (USB 3.0) connector of 3.2  Gbit/s (400  MB/s) after accounting for encoding overhead. In addition, we must perform about 400,000 autocorrelation operations (192×128  pixels, 16-lags, see Sec. 3) in each integration period (tens of ms) in real time. Such large data rates and compute-intensive operations are not sustainable with standard protocol, thus data compression on camera is needed, using FPGA, before passing the processed data to PC.

We synthesized a total of 128 ACs on the FPGA. Each AC was composed of 16 linearly spaced lags with a minimum time lag (τmin) of 38.4  μs and a maximum time lag (τmax) of 614.4  μs, related by τmax=16τmin. The sensor outputs the photon count for each τmin period, which is equivalent to the sensor exposure time. Thus the shortest τmin is limited by the maximum frame rate of the sensor of 26 kfps. Although we can, in principle, adjust τmin by increasing the exposure time forcing a lower frame rate, the longest τmin is capped by the pixel saturation value of 255 counts limited by the in-pixel counter.

Due to the limited memory resources on the chosen FPGA, only 64 columns of pixels could be stored in memory over a g2(τ) integration period (Tint) of 60 ms. The FPGA can therefore compute 12,288 ACFs from the group of 192×64  pixels in the SPAD array. This is half of the full array in our 192×128 SPAD array.

The ACF as a function of time lag τ for each pixel i can be estimated by computing the following equation:

Eq. (1)

g2i(τ)=ni(t)ni(t+τ)ni(t)2,
where n(t) is the number of photons recorded in time bin t, τ is the time lag, and the square bracket denotes the average over the integration time Tint. The denominator component is the square of the average photon count over Tint. The variation in the denominator value across the time lags (16 lags) is negligible at long Tint of 60 ms. We then calculated the ensemble average of the g2 curves measured over M pixels as

Eq. (2)

g¯2(τ)|M=1Mi=1Mg2i(τ).

We measured g2 with Tint=60  ms and repeated the measurement for up to 6 s, contiguously with no spacing in between, for a total of 100 g2 measurements per pixel. We calculated the time average MEAN[g2(τ)] and STD[g2(τ)] of g2 over all of the 100 integration periods. We define the SNR of the ensemble average as SNR[g¯2(τ)]=MEAN[g¯2(τ)1]/STD[g¯2(τ)].

Equation (1) was implemented on the FPGA for each pixel i with the following equation:

Eq. (3)

g2i(τ)=N·(k=1Nni(k)ni·(kfτ))(k=1Nni(k))2,
where k is the frame number and N is the total number of frames in an ACF integration period. The above equation must be computed for each of the fτ=1 to 16 frame lags.

The ACs were designed on FPGA to compute Eq. (3). The 128 ACs are grouped into pairs for a total of 64 blocks as shown in Fig. 2. Each of the 64 column-parallel AC blocks contains two sets of 8b, 16-lag ACs and two sets of 4 normalization dividers. Each set processes a half column of 96 pixels, one set for the top half of the array and one set for bottom half of the array. The column-parallel AC block is an instantiable module that can be replicated and easily connected to the rest of the surrounding logic. It is therefore simple to scale the number of column-parallel ACs from 64 to 128 blocks to serve the remaining unused columns of the SPAD array. Currently, this is a limitation of the chosen FPGA, but a larger off-the-shelf FPGAs would not impose such limit. The AC Control FSM controls the execution of the operations within a column-parallel AC block. The FSM is replicated for each column-parallel AC block as having it local to each block allowed better convergence of timing as opposed to routing control signal from a single shared FSM to all the 64 blocks. A g2 column adder and sum memory aggregate all per-pixel ACFs in a column. Finally, the global FSM, g2 global adder, and barrel shifter sum the g2 ACFs and take the average of columns into a single ensemble g2 ACF for the array of 192×64  pixels.

Fig. 2

(a) Diagram of the column arrangement of the QuantICAM SPAD camera with the respect to the assignment to the column-parallel AC blocks. (b) The block diagram of each of the 64 column-parallel AC blocks.

JBO_28_5_057001_f002.png

The block diagram of one of the 128 8b, 16-lag ACs is shown in Fig. 3(a). The AC is made up of an 8b, 16-element register (SRreg) to store the photon count of 16 consecutive frames. A barrel shifter is used to shift the SRreg by the new 8b frame count value n(k) and write the shifted SRreg back into the AC memory, acting as a virtual shift register for the 16 lag frames. A dual-port AC memory, as shown in Fig. 3(a), is used to store the value of the shift register and AC components for each lag over the integration period of N frames for all pixels in a half-column. In order to store the 8b shift register for 16 lags, four 32b memory addresses are required. The allocation of the g2 components to the addresses in the AC memory is shown in Fig. 3(b).

Fig. 3

(a) Diagram of the blocks within an 8b, 16-lag AC. (b) Arrangement of data within the AC memory in each AC.

JBO_28_5_057001_f003.png

Each of the 16 lags requires a multiplier to calculate the product of n(k) with each of the 16 previous frame counts n(k+τ). The products n(k)n(k+τ) are accumulated for each lag over an integration period of N frames and stored in the AC memory for each lag τ. These memory locations are known as product accumulators (PA) and there are 16 for each pixel. As multiply and accumulate cores (MACs) in the FPGA are limited, a single multiplier and integrator are shared among the 16 lags of each column AC. The FSM multiplexes the MAC for each lag.

To normalize each PA, the sum of the photons over all frames in an integration period is stored in a memory address for normalization at the end of the N frames. This AC component is known as the photon count accumulator (PCA). Each pixel requires 21 memory locations: 1 for the PCA, 4 for the shift register, and 16 for the PAs. In total to store the 96 pixels in a column, the AC memory is 32b by 2016 addresses.

The FSM controls the timing of the shift register update, multiplexing of the multiplication and accumulation of the all PAs and the integration of the PCA. The same FSM controls the reading and writing to the AC memory, computes and updates all memory locations for a pixel within a row-time to guarantee a frame rate of 26 kfps. The operations of the 8b, 16-lag AC at each new frame is summarized in Table 1. The AC takes 42 clock cycles of the 100-MHz system clock to update a pixel at each new frame.

Table 1

AC frame update operations.

Clock cycleOperation description
 1• Read PCA value from AC memory
 2• Sum n(k) to read PCA value and write updated PCA value to AC memory
 3• Read SR lags 1 to 4 addresses and store SR lags in local register SRreg[0:3]
 4• Write to SR lags 1 to 4 addresses to update SR shifted with n(k) as first element of SR
 5• Read SR lags 5 to 8 addresses and store SR lags in local register SRreg[4:7]
 6• Write to SR lags 5 to 8 addresses to update SR shifted by one frame count
 7• Read SR lags 9 to 12 addresses and store SR lags in local register SRreg[8:11]
 8• Write to SR 9 to 12 addresses to update SR shifted by one frame count
 9• Read SR elements 13 to 16 addresses and store SR lags in local register SRreg[12:15]
10• Write to SR 13 to 16 addresses to update SR shifted by one frame count
11• Read PA for lag fτ=1
12• MAC: PAnew(1)=SRreg[0]·n(k)+PAprev and write MAC result in PA address for lag fτ=1
13-42• Read PA for lag fτ=216

• MAC: PAnew(fτ)=SRreg[fτ1]·n(k)+PAprev and write MAC result in PA address for lag fτ=216

To normalize the ACF, the PA values must be multiplied by the number of frames N in the integration period and divided by the PCA squared [see Eq. (3)]. At the end of the integration period, the contents of the AC memory are sent to the normalization dividers, which compute the square of the PCA and divide each lag PA by PCA2. The normalization operation of all pixels is completed in <2 frame times. The normalization starts as soon as the first pixel in the half-column has completed the last of the N frames and while the other pixels are finishing and updating the memory for the N’th frame. This ensures the normalization is completed before the pixels have to address the memory to start integrating again over a new N-frame integration period. As the division operation is a pipelined operation taking 20 clock cycles to compete for a single lag normalization, 4 dividers are used to compute 4 lags in parallel to fulfill the timing requirement of completing the normalization of all pixels in <2 frame times. This ensures that the normalization operation does not introduce a dead time in the acquisition to elaborate the data. The dual-port memory was chosen over the single-port option so that the integration and operations on the AC memory for the N’th frame could run concurrent to the dividers reading out and elaborating the pixel addresses that have already completed the AC operation for the N’th frame.

The normalized per-pixel ACFs can be optionally streamed out as they are being calculated by the column-parallel normalization dividers. This is a debug operation modality that allows the user to probe the individual per-pixel ACFs and monitor how these contribute to the SNR scaling. In the default modality, the eight dividers send the normalized per pixel ACFs to the g2 column adder [see Fig. 2(b)] which sums all the per-pixel ACFs of a column into the g2 column memory as each new pixel from the column is normalized. Once all pixels from a column are normalized and their ACFs summed and stored into the g2 column memory, the global g2 adder takes the sum of the ACFs from all columns. The final barrel shifter divides the result by 64 using a binary shift to normalize to the 64 columns. Division by the number of pixels in a column (192) and multiplication by the number of frames in the integration period N can be completed in software as it is a constant coefficient factor applied to the final ensemble g2 result.

3.

Results and Discussion

To demonstrate the capability of our large SPAD array with on-FPGA processing for increased SNR, we performed a rotating diffuser experiment. Figure 4(a) shows the experimental setup, consisting of a long coherence length (∼9 m) continuous wave laser emitting at 785 nm wavelength, a rotating diffuser plate, and a SPAD camera. Here we used the 192×128 SPAD camera (QuantICAM), but only half of the array was read out to FPGA as discussed above. The laser output was coupled to the phantom using a single-mode fiber. The diffused photons were collected by a multimode fiber (MMF) in a transmission geometry and coupled to the SPAD camera. The fiber-SPAD distance (z) was adjusted using a translating lens tube to calibrate the diameter of the speckles such that the average diameter of the speckles is at least twice larger than the SPAD active area (Nyquist).12 The SPAD camera consisted of 24,576 SPADs arranged in a 192×128 array, with a pixel pitch of 18.4  μm×9.2  μm and a 13% active fill factor. In comparison, the SPAD camera used in prior work was a 32×32 array, with a pixel pitch of 50  μm×50  μm and a 1.5% active fill factor.6,7

Fig. 4

(a) Experimental setup for multispeckle DCS using a rotating diffuser phantom. (b) Ensemble averaged intensity ACF g2(τ) of the photon counts from the rotating diffuser phantom. The measured g2(τ) can be fitted well with g2(τ)1=βexp(τ2/τc2), where β (=0.42) is the coherence factor, τ is the time delay in unit of the exposure time Tex=38.4  μs, and τc (=7.2) is the correlation time. After converting the time delay unit, we obtained an estimated correlation time of τc=278  μs. (c) The speed of the motorized rotation stage (K10CR1, Thorlabs) was controlled by a user interface in PC. The diffuser plate (Biomimic™ optical phantom, INO) has the absorption coefficient μa=0.1  cm1 and reduced scattering μs=13.4  cm1, and it was cut into a cylindrical shape with dimension of 1 inch in diameter and 5-mm-thick. The source and detector fibers have their axes aligned, but offset from the center of the rotation by about 9 mm.

JBO_28_5_057001_f004.png

In the limit of low photon count rates (shot-noise limited regime), the SNR of g¯2(τ0) is proportional to Nph×Tint×M, where Nph is the detected photon count rate, Tint is the integration time, and M is the number or pixels.6,13 Hence, in addition to the 24× larger pixel count, our SPAD camera offers slightly more than 8× larger active fill factor that allows a single MMF to couple the full-size SPAD array with independent speckle mode per pixel. To measure the speckle turnover time trace, we mounted the diffuser plate on a motorized rotation stage. The diffuser plate (Biomimic™ optical phantom, INO) has the absorption coefficient μa=0.1  cm1 and reduced scattering coefficient μs=13.4  cm1. Figure 4(b) shows the measured intensity ACF g2(τ) of the photon counts, where the time delay τ is in the unit of the exposure time Tex=38.4  μs, with rolling shutter. The transmitted light was collected using MMF with 1000  μm core diameter.

To validate the SNR gain [defined as the ratio SNR[g¯2(τ)]/SNR[g2i(τ)] for the first time bin], we compared the STD of the single-pixel g2i and the ensemble averaged g¯2. Figure 5 shows separate rotating diffuser measurements at an angular speed of 12 deg/s, with SPAD exposure time Tex=61.44  μs. The solid line in Fig. 5(a) (g2i) and 5b (g¯2) corresponds to the mean value of g2(τ). The shade corresponds to ±2 STD with respect to the mean value. The STD of g¯2(τ=Tex) [Fig. 3(b)] is about 110 times smaller than that of g2i(τ=Tex) [Fig. 3(a)], which is much smaller than the size of the symbols in the plot. Figure 5(c) shows the SNR of g¯2(τ) at increasing size of the pixel ensemble M of up to 12,288. Averaging over a larger ensemble of pixels leads to higher SNR. Figure 5(d) shows the SNR gain of g¯2(τ=Tex) as a function of the size of the pixel ensemble M of up to 12,288, normalized to that of the single-pixel. The measured SNR gain (solid circles) increases as the square root of the ensemble size (solid line), as predicted by the multispeckle DCS model.6

Fig. 5

(a) Time statistics of the pixel intensity ACF g2i(τ). (b) Time statistics of the ensemble intensity ACF g¯2(τ). (c) SNR of g¯2(τ) for increasing size of the pixel ensemble M=1 to 12,288. (d) SNR gain of g¯2(τ=Tex) as a function of the size of the pixel ensemble M=1 to 12,288. The exposure time was Tex=61.44  μs.

JBO_28_5_057001_f005.png

4.

Conclusion

We have demonstrated an SNR gain of 110 with respect to the single-pixel multispeckle DCS using half of the 192×128 SPAD array, with pixel active fill factor of 13%. This is an order of magnitude larger pixel count than the 32×32 SPAD array with fill factor of 1.5% in prior works.6,7 In addition to the large pixel count, we have also demonstrated a real-time data compression scheme implemented in the readout of the SPAD camera by calculating the ensemble g2 ACF entirely on FPGA, showing an improvement of over an order of magnitude in the number of parallel channels that can be processed in real time. Shifting the g2 calculations from a high-performance workstation computer to FPGA has enabled the running of real-time multispeckle DCS measurements using a standard PC, e.g., a laptop. This effort should help democratize the use of large-pixel-count SPAD cameras for multispeckle DCS in broad biomedical optics research. Although the FPGA-embedded processing plays a crucial role to jump start large array multispeckle DCS research due to its reconfigurability as we explore different operation modes (e.g., photon-counting, DCS, and time-domain DCS), our complete on-FPGA processing chain demonstrates that the digital circuit resources required are compatible with embedded on-chip-based processing. Thus we envision that ultimately real-time data compression is achievable in an advanced computational SPAD camera based with integrated focal-plane autocorrelators.

Finally, recent theoretical investigation by Cheng et al.14 suggested that to be sensitive to local, subsecond latency hemodynamic response of the cerebral cortex at 15  mm beneath the scalp surface, we need a shorter exposure time that can resolve the ACF with a time constant of the order of tens of μs and a larger pixel count of the order of 104 to 106  pixels. Although our results with on-FPGA processing offer great promise for multispeckle DCS, currently the shortest exposure time of our SPAD camera is limited to Tex=38.4  μs and the photon detection probability of the pixel is 8% at an excess bias of 1.7 V and wavelength of 785 nm.11,15 Despite this exposure time limitation, we believe that the QuantICAM DCS system can benefit real-time applications, such as neuroscience for smaller animals with skull/scalp thickness much smaller than that of humans which leads to longer correlation time, thus can be measured using longer exposure time from QuantICAM SPAD camera. Other examples include measurements on human extracerebral blood flow and human peripheral blood flow (e.g., wrist) for bedside health monitoring. Another example includes regression from extracerebral blood flow using short source–detector separation that leads to correlation time of the order of tens of μs.

Intense research effort in SPAD camera development is currently ongoing to improve SPAD camera capabilities toward even larger pixel count (106), shorter exposure time (sub-μs), and higher detection probability.1618 Soon, we should expect high-performance SPAD cameras with FPGA-embedded or even on-chip computing that could surpass the multispeckle DCS requirements for noninvasive detection of local brain activation.

Disclosures

The authors declare no potential conflicts of interest.

Code, Data, and Materials Availability

The data and code (FPGA Verilog RTL and MATLAB software for data elaboration) that support the findings of this study are available from the corresponding author on reasonable request.

Acknowledgments

We would like to thank Peter Bracewell and E-Fann Saung for their engineering support in QuantICAM camera bringup, and Stephanie Naufel and Mark Chevillet for their support in setting up the research collaboration. This research was supported by funding from Meta Platforms Inc.

References

1. 

D. A. Boas et al., “Establishing the diffuse correlation spectroscopy signal relationship with blood flow,” Neurophotonics, 3 031412 https://doi.org/10.1117/1.NPh.3.3.031412 (2016). Google Scholar

2. 

T. Durduran et al., “Diffuse optical measurement of blood flow, blood oxygenation, and metabolism in a human brain during sensorimotor cortex activation,” Opt. Lett., 29 1766 https://doi.org/10.1364/OL.29.001766 OPLEDP 0146-9592 (2004). Google Scholar

3. 

T. Durduran and A. G. Yodh, “Diffuse correlation spectroscopy for non-invasive, microvascular cerebral blood flow measurement,” Neuroimage, 85 51 https://doi.org/10.1016/j.neuroimage.2013.06.017 NEIMEF 1053-8119 (2014). Google Scholar

4. 

E. M. Buckley et al., “Diffuse correlation spectroscopy for measurement of cerebral blood flow: future prospects,” Neurophotonics, 1 011009 https://doi.org/10.1117/1.NPh.1.1.011009 (2014). Google Scholar

5. 

H. Ayaz et al., “Optical imaging and spectroscopy for the study of the human brain: status report,” Neurophotonics, 9 S24001 https://doi.org/10.1117/1.NPh.9.S2.S24001 (2022). Google Scholar

6. 

E. J. Sie et al., “High-sensitivity multispeckle diffuse correlation spectroscopy,” Neurophotonics, 7 035010 https://doi.org/10.1117/1.NPh.7.3.035010 (2020). Google Scholar

7. 

W. Liu et al., “Fast and sensitive diffuse correlation spectroscopy with highly parallelized single photon detection,” APL Photonics, 6 026106 https://doi.org/10.1063/5.0031225 (2021). Google Scholar

8. 

G. Dietsche et al, “Fiber-based multispeckle detection for time-resolved diffusing-wave spectroscopy: characterization and application to blood flow detection in deep tissue,” Appl. Opt., 46 8506 https://doi.org/10.1364/AO.46.008506 APOPAI 0003-6935 (2007). Google Scholar

9. 

J. D. Johansson et al., “A multipixel diffuse correlation spectroscopy system based on a single photon avalanche diode array,” J. Biophotonics, 12 1 https://doi.org/10.1002/jbio.201900091 (2019). Google Scholar

10. 

J. Buchholz et al., “FPGA implementation of a 32 × 32 autocorrelator array for analysis of fast image series,” Opt. Express, 20 17767 https://doi.org/10.1364/OE.20.017767 OPEXFF 1094-4087 (2012). Google Scholar

11. 

R. K. Henderson et al., “A 192 × 128 time correlated SPAD image sensor in 40-nm CMOS technology,” IEEE J. Solid-State Circ., 54 1907 https://doi.org/10.1109/JSSC.2019.2905163 IJSCBC 0018-9200 (2019). Google Scholar

12. 

S. J. Kirkpatrick, D. D. Duncan and E. M. Wells-Gray, “Detrimental effects of speckle-pixel size matching in laser speckle contrast imaging,” Opt. Lett., 33 2886 https://doi.org/10.1364/OL.33.002886 OPLEDP 0146-9592 (2008). Google Scholar

13. 

C. Zhou et al., “Diffuse optical correlation tomography of cerebral blood flow during cortical spreading depression in rat brain,” Opt. Express, 14 1125 https://doi.org/10.1364/OE.14.001125 OPEXFF 1094-4087 (2006). Google Scholar

14. 

X. Cheng et al., “Measuring neuronal activity with diffuse correlation spectroscopy: a theoretical investigation,” Neurophotonics, 8 035004 https://doi.org/10.1117/1.NPh.8.3.035004 (2021). Google Scholar

15. 

S. Pellegrini et al., “Industrialised SPAD in 40 nm technology,” in IEEE Int. Electron. Devices Meeting (IEDM), 16.5.1 –16.5.4 (2017). https://doi.org/10.1109/IEDM.2017.8268404 Google Scholar

16. 

C. Bruschini et al., “Single-photon avalanche diode imagers in biophotonics: review and outlook,” Light Sci. Appl., 8 87 https://doi.org/10.1038/s41377-019-0191-5 (2019). Google Scholar

17. 

K. Morimoto et al., “A megapixel time-gated SPAD image sensor for 2D and 3D imaging applications,” Optica, 7 346 https://doi.org/10.1364/OPTICA.386574 (2020). Google Scholar

18. 

M. Wayne et al., “A 500 × 500 dual-gate SPAD imager with 100% temporal aperture and 1 ns minimum gate length for FLIM and phasor imaging applications,” IEEE Trans. Electron. Devices, 69 2865 https://doi.org/10.1109/TED.2022.3168249 IETDAI 0018-9383 (2022). Google Scholar

Biography

Francescopaolo Mattioli Della Rocca received his MEng and PhD degrees in electronic and electrical engineering from The University of Edinburgh, Edinburgh, UK, in 2015, focusing on single-photon avalanche diode CMOS image sensors. He is now at Europe Technology Development Centre, Sony Semiconductor Solutions, Trento, Italy.

Edbert J. Sie received his PhD in physics from Massachusetts Institute of Technology (MIT) in 2017 and a postdoctoral fellowship from Stanford University, where he led research in ultrafast optics and imaging techniques, TOF detection, quantum materials, and atomically thin semiconductors. He is an optical systems scientist at Reality Labs Research of Meta, developing optical sensors for brain–computer interface and AR/VR applications.

Ryan Catoen is a software engineer at Reality Labs Research of Meta. Previously, he studied computer science at the University of Waterloo, Canada. His research interests focus on accelerating the development of next-generation noninvasive brain–computer interfaces.

Francesco Marsili received his PhD in physics from the Ecole Polytechnique Federale de Lausanne, Switzerland, in 2009. He is a research scientist manager at Reality Labs Research of Meta, developing diffuse optical techniques for the measurement of biosignals. Previously, he led research in single-photon detectors, optical communication, quantum optics, superconductivity, and nanofabrication at the Jet Propulsion Laboratory, the National Institute of Standards and Technology, and MIT.

Robert K. Henderson obtained his PhD in 1990 from the University of Glasgow. He is a professor at the School of Engineering of The University of Edinburgh. From 1991, he was a research engineer at Swiss Centre for Microelectronics, Neuchatel, Switzerland. In 1996, he was appointed as a senior VLSI engineer at VLSI Vision Ltd., Edinburgh, UK, where he worked on the world’s first single chip video camera. From 2000, as a principal VLSI engineer in the STMicroelectronics Imaging Division, he developed image sensors for mobile phone applications. He joined Edinburgh University in 2005, designing the first SPAD image sensors in nanometer CMOS technologies in MegaFrame and SPADnet EU projects. In 2014, he was awarded a prestigious ERC advanced fellowship. He is a fellow of the Royal Society of Edinburgh and the Institute of Electrical and Electronics Engineers.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Francesco Mattioli Della Rocca, Edbert J. Sie, Ryan Catoen, Francesco Marsili, and Robert K. Henderson "Field programmable gate array compression for large array multispeckle diffuse correlation spectroscopy," Journal of Biomedical Optics 28(5), 057001 (8 May 2023). https://doi.org/10.1117/1.JBO.28.5.057001
Received: 19 December 2022; Accepted: 24 April 2023; Published: 8 May 2023
Lens.org Logo
CITATIONS
Cited by 3 scholarly publications.
Advertisement
Advertisement
KEYWORDS
Single photon avalanche diodes

Cameras

Field programmable gate arrays

Signal to noise ratio

Photons

Actinium

Autocorrelation

RELATED CONTENT


Back to Top