Skip to main content
Science Advances logoLink to Science Advances
. 2025 Jun 11;11(24):eadt3558. doi: 10.1126/sciadv.adt3558

RF-photonic deep learning processor with Shannon-limited data movement

Ronald Davis III 1,*, Zaijun Chen 1,2, Ryan Hamerly 1,3, Dirk Englund 1
PMCID: PMC12153973  PMID: 40498817

Abstract

Edholm’s law predicts exponential growth in data rate and spectrum bandwidth for communications. Owing to exponentially increasing deep neural network computing demands and the slowing of Moore’s law, new computing paradigms are required for future advanced communications like 6G. Optical neural networks (ONNs) are promising accelerators but struggle with scalability and system overhead. Here, we introduce our multiplicative analog frequency transform optical neural network (MAFT-ONN), an artificial intelligence hardware accelerator that experimentally computes fully analog deep learning on raw radio frequency (RF) signals, performing modulation classification that quickly converges to 95% accuracy. MAFT-ONN also exhibits scalability with nearly 4 million fully analog operations for MNIST digit classification. Because of the Shannon capacity–limited analog data movement, MAFT-ONN is also hundreds of times faster than traditional RF receivers.


MAFT-ONN is a photonic AI hardware accelerator that performs fully analog deep learning on raw RF signals.

INTRODUCTION

Artificial intelligence (AI) has been revolutionizing a broad range of fields, including radio frequency (RF) signal processing and advanced communications. In environments where the spectrum is congested with several users and many channels, hand-engineered systems are becoming increasingly infeasible. Here, AI can be leveraged to process the increasingly complex spectral environment while meeting the rising demand for higher wireless data rates. Signal processing AI enables the next generation of communications such as 6G, where capabilities like cognitive radio, fingerprinting, and dynamic resource allocation may take pivotal roles (13).

For RF signal processing, state-of-the-art AI approaches first digitize the IQ data and then either convert the signal into an N × 2 IQ tensor (4, 5) or compute the spectrogram (6, 7) (or other time-frequency transform) to convert the signal into an image. The preprocessed signal is then inserted into a convolutional neural network (CNN) or other deep learning model for tasks like signal classification or device fingerprinting.

However, while digital processors can compute CNNs with high accuracy, these methods introduce substantial latency that make real-time spectrum processing impossible for digital processors. This is because state-of-the-art digital architectures require several steps to move the large volume of RF data to and from the compute. The alternative is to store the data for later offline analysis, but this is not feasible for time-sensitive tasks. In addition, these current approaches struggle to maintain high performance while keeping cost, size, weight, and power low due to the need to use high-performance devices like FPGAs, GPUs, and RFSoCs (810).

Optical systems promise AI acceleration by encoding, routing, and processing analog signals in optical fields, allowing for operation at the quantum noise limit with high bandwidth and low energy consumption. Optical neural network (ONN) schemes rely on (i) performing linear algebra intrinsically in the physics of optical components and/or (ii) in-line nonlinear transformations. For (i), past approaches include Mach-Zehnder interferometer meshes (1115), on-chip micro-ring resonators (1619), wavelength-division multiplexing (2022), photoelectric multiplication (23), spatial light modulation (2429), optical scattering (30), optical attenuation (31), vertical-cavity surface-emitting lasers (32), and optical diffraction (3336). For (ii), past approaches include optical-electrical-optical elements (14, 31, 3739) and all-optical (18, 28, 4043) approaches. However, to fully take advantage of the potential ultralow latency and energy consumption available in photonics, it is necessary to implement linear and nonlinear operations together with minimal overhead. Simultaneously achieving (i) and (ii) in a way that preserves high hardware scalability and performance has been an open challenge.

Our multiplicative analog frequency transform optical neural network (MAFT-ONN) architecture simultaneously achieves (i) and (ii) for deep neural network (DNN) inference with high scalability in both DNN size and layer depth. We experimentally demonstrate the MAFT-ONN in a three-layer DNN for inference of Modified National Institute of Standards and Technology (MNIST) images and modulation classification. In this architecture, we encode neuron values in the amplitude and phase of frequency modes, and “photoelectric multiplication” (23) performs matrix-vector products in a single shot. The nonlinear activation for each layer is achieved using the nonlinear region of an electro-optic modulator, thus enabling a scalable front-to-back photonic hardware accelerator for DNNs. Figure 1A contextualizes use cases for the MAFT-ONN processor.

Fig. 1. An overview of the MAFT-ONN architecture.

Fig. 1.

(A) The MAFT-ONN processor accelerates both traditional signal processing operations and AI inference for waveforms like radio waves. The analog received waveform is fed into MAFT-ONN for fully analog processing, after which the output may be read out digitally using an analog-to-digital converter (ADC) or fed into another analog system. (B) An outline of the MAFT-ONN architecture. Each photoelectric multiplication physically computes either a fully connected (FC) or a 1D convolution (CONV) layer. The nonlinear activation (NL) for each layer physically corresponds to the nonlinear region of the following modulator. The units can be cascaded to implement several DNN layers fully in analog with no digital overhead. (C) A close-up of a single FC or CONV layer. For an FC layer, the weight signal is programmed such that the photoelectric multiplication yields a matrix-vector product in the frequency domain, where the green region in the frequency domain is isolated with a filter. For a CONV layer, all frequencies of the output signal are used. S. Wilcox (MIT) aided in designing the figure.

MAFT-ONN ARCHITECTURE

As illustrated in Fig. 1B, a series of DNN layers corresponds to a cascading of photoelectric multiplications where each one computes either a fully connected (FC) or 1D convolution (CONV) layer. The nonlinear activation (NL) for all neurons in a given layer is achieved by operating in the nonlinear regime of the following modulator.

Figure 1C details an experimental example of a 2 × 2 matrix-vector product using the MAFT scheme. In practice, the electrical voltage signal VX(t) is the incoming RF waveform and VW(t) is the weight signal generated by MAFT-ONN. With an underbar denoting the complex representation of the signal, the terms Re[V_(t)] and Im[V_(t)] represent a voltage signal and its 90 phase-shifted copy, the combination of which is required to achieve single-sideband suppressed carrier (SSB-SC) modulation. (See Methods for experimental details.)

The SSB-SC modulated input and weight signals are then photoelectrically multiplied to yield the output voltage signal: Vout(t)Im[V_X*(t)V_W(t)] . The output signal’s Fourier Transform V~out(ω) corresponds to the desired matrix multiplication performed in a single shot. This matrix multiplication is achieved by appropriately programming the frequency content of the weight signal VW(t) . For an FC layer, the frequencies of V~out(ω) that correspond to the matrix product are within the green region in Fig. 1C and are isolated using a bandpass filter. For a CONV layer, all frequencies remain. See section SG for the generalized matrix multiplication algorithm.

RESULTS

MNIST digit inference

We experimentally demonstrated a proof of concept of the MAFT-ONN architecture for a three-layer DNN trained to classify MNIST digits. As shown in Fig. 2A, the DNN consists of two CONV weight signals with a nonlinear activation for the hidden layer. The frequency-encoded input layer consists of a flattened 14 × 14 MNIST image and thus contained 196 frequencies that represent the neurons. This was convolved with a weight CONV kernel of 19,600 frequencies to yield the hidden layer of 39,100 neurons (frequencies). After using a dual-parallel Mach-Zehnder modulator (DPMZM) as the nonlinear activation, the hidden layer was next convolved with a second weight CONV kernel of 1000 frequencies to yield the output layer of 10 neurons, one output neuron for each of the MNIST digits.

Fig. 2. Experimental demonstration of MAFT-ONN.

Fig. 2.

(A) An example of an experimental inference of a 14 × 14 MNIST image encoded into the input signal VX(1)(t) . The two convolutions are achieved with the weight signals VW(1)(t) and VW(2)(t) , respectively. The nonlinear activation is achieved using an amplifier so that the signal Vout(1)(t) reaches the nonlinear region of the next DPMZM. As shown in the zoom plot of Vout(2)(t) , the image is correctly classified. (B) In the upper half, 2D histograms compare the experimental output values Y^ to the expected curve fitted value Y . In the lower half, 1D histograms plot the error YY^ . The scalar-scalar plot contains 10,000 randomized 1 × 1 matrix products, yielding 9-bit precision compared to the curve fit. The matrix-vector plot contains 10,000 randomized 10 × 10 matrix products (thus 100,000 values), yielding 8-bit precision. (C) An experimental characterization of the nonlinear activation function of an MZM. We programmed VX(1)(t) as a 10 × 1 input vector, and gradually increased its amplitude until it reached the nonlinear regime of the MZM. We then curve fitted an analytical model to the experimental data. (D) A confusion matrix of the experimental three-layer DNN over 10,000 14 × 14 MNIST images, yielding an experimental accuracy of 86.85%.

The number of multiply-and-accumulates (MACs) computed per MNIST image inference is the sum of the MACs in two CONV layers. The number of MACs computed for a 1D convolution between two vectors with lengths N1 and N2 is simply N1N2 . This is because the number of MACs when the two vectors fully overlap (assuming N2>N1 ) is (N2N1+1)N1 and the number of MACs when the convolution is at the edges is 2i=1N11i=(N11)N1 . Therefore, the number of MACs experimentally computed per MNIST inference is: (14·14)·19,600+1000·10=3,851,600 MACs. The first term is from convolving the flattened input image with the weights and the second term only counts the MACs used for the output neurons.

The three-layer experimental DNN inferred 10,000 14 × 14 MNIST images, where the digital DNN has an accuracy of 92.52% and the experimental DNN has an accuracy of 86.85%. One contribution to the experimental inaccuracy is ripples found in the experimental nonlinear activation function, perhaps due to the path length difference of the interferometer (see section SC for data on this). Other methods to increase the accuracy include introducing voltage drivers to the DPMZMs to increase the SNR by an order of magnitude (see Methods) and performing the DNN training in situ on the hardware itself (44). The confusion matrix of the experimental DNN is shown in Fig. 2D.

LTI signal processing

The MAFT architecture is capable of computing arbitrary finite impulse response (FIR) linear time invariant (LTI) operations in the frequency domain. When interpreting the frequency modes as discrete LTI samples, the weight signal VW(t) can be programmed to implement arbitrary frequency-domain FIR LTI convolutions with the input signal VX(t).

Let the input voltage signal be VX(t)=n=1NXncos(n·Δω·t) and the LTI filter signal be VW(t)=r=1RWrcos(r·Δω·t) . Then, the frequency-domain LTI interpretation is: x[n]XnnΔω and w[n]WnnΔω . Hence, the values of x[n] and w[n] are mapped to the frequency mode at nΔω . With this interpretation, appropriately programming the SSB-SC conditions of the DPMZMs enables MAFT-ONN to compute convolutions that correspond to y[n]=(x**w)[n]=k=x*[k]w[nk] . Therefore, MAFT yields an LTI convolution for real-valued signals and a still useful LTI-like operation for complex-valued signals. See section SN for the mathematical derivation of the LTI framework from physics principles. Figure 3 (A to C) illustrates experimental results of various signal processing operations.

Fig. 3. Experimental demonstrations of the signal processing capabilities of MAFT-ONN.

Fig. 3.

The signal processing operations in (A) to (C) used the experimental setup in Fig. 1C. (A) The frequency-domain LTI framework was used to implement a Wiener filter that recovers a signal that suffered from frequency–cross-talk and AWGN. (B) The matrix-vector product method from Fig. 1C was used to implement an LLSE estimator to recover a signal that suffered jointly Gaussian noise in various frequency bands. (C) The frequency-domain LTI framework was used to scan the spectrum of a signal for a specific frequency signature. (D) The three-layer DNN hardware in Fig. 2A was used to experimentally implement modulation classification on raw RF signals using MAFT-ONN. For single-shot inference, the modulation classification experimentally achieved 85.0% accuracy compared to the digital accuracy of 89.2%. (E) With the experimental single-shot accuracy as the baseline, we show that a majority vote multimeasurement scheme will asymptotically improve the accuracy to 100%. With only five measurements, the accuracy improves to 95%. B-FM, broadcast frequency modulation; CPFSK, continuous phase frequency shift keying; DSB-AM, double sideband amplitude modulation; GFSK, Gaussian frequency shift keying; SSB-AM, single sideband amplitude modulation.

FIR Wiener filter

Let VX(t) be a frequency channelized signal with 1000 frequency modes where the amplitude of each frequency encodes a symbol. Here, we assume a 4-bit encoding scheme where each symbol is equiprobable a priori. After VX(t) is generated, cross-talk is introduced between each of the frequency modes and then additive white Gaussian noise (AWGN) is added to yield the distorted signal VY(t) . The goal is to estimate the original signal VX(t) from the received signal VY(t).

The frequency content of VX(t) and VY(t) is modeled as jointly wide-sense stationary (WSS) random processes x[n] and y[n] , respectively. Given that the frequency cross-talk and AWGN are known, we calculate the covariance matrices to construct an FIR Wiener filter h[n] of length 50, meaning that the previous 50 samples of y[n] will be used to estimate the next sample. The estimated signal x^[n] is then computed by applying the Wiener filter to yield: x^[n]=(y*h)[n].

The effectiveness of the Wiener filter is evaluated using the mean squared error (MSE) between the original signal and the estimated signal. The MSE is calculated over 10 experimental measurements of the instance of the random process. The experimental MSE of the uncorrected signal xy2 has a mean of 39.46 mV2 and an SD of 0.27mV2 , and the experimental MSE of the corrected signal xx^2 has a mean of 24.59 mV2 and an SD of 0.14 mV2 , yielding an average MSE improvement of 37.69%. This closely matches the theoretical MSE improvement of 38.78%. Figure 3A plots one of the measurements for each signal.

According to Parseval’s theorem, estimating VX(t) in the frequency domain (opposed to the time domain) is effective because the energy of the error signal is the same in both the time and frequency domains. Hence, given the appropriate context, the MAFT frequency-domain estimation is a practical tool for signal processing scenarios.

LLSE estimator

The linear least squares estimator (LLSE) is more general than the Wiener filter and works on non-WSS random processes. Let VX(t) be frequency channelized as in the previous example, except this time it is split up into three frequency bands each with different amplitude encoding characteristics. In frequency-domain LTI notation, VX(t) is broken into its three frequency bands as x[n]=xb1[n]+xb2[n]+xb3[n].

Each of the three bands goes through a combination of shared and individual sources of noise, which is modeled as jointly Gaussian noise, to yield the distorted signal y[n] . Because each of the frequency bands have different encoding schemes and different sources of noise, x[n] and y[n] are no longer WSS and cannot be estimated using a Wiener filter.

Assuming we have knowledge of the jointly Gaussian noise, we constructed the optimal LLSE filter by calculating the appropriate covariance matrices. We use the scheme to create the filter h[n] that implements the LLSE estimator to produce x^[n].

As before, 10 experiments of the instance were measured to calculate the experimental MSE. The experimental uncorrected MSE xy2 has a mean of 479.58 mV2 with an SD of 1.03 mV2 , and the experimental corrected MSE xx^2 has a mean of 193.38mV2 and an SD of 22.98 mV2for an average MSE improvement of 59.68%. This, again, is close to the theoretical MSE improvement of 63.17%. Figure 3B plots one of the measurements for each signal.

Matched filters

A useful benefit of MAFT is that a single photoelectric multiplication can be used to simultaneously scan the entire RF spectrum for a target signal. Let Vi(t) be a series of signals that we wish to scan for a specific frequency signature V~target(ω) . The frequency-domain LTI representations of Vi(t) and V~target(ω) are xi[n] and xtarget[n] , respectively.

Figure 3C shows three examples of applying the matched filter to various received signals. For x1[n] , the target xtarget[n] is located lower in the spectrum, and thus, a peak appears there in the matched filter output. For x2[n], the target is located higher in the spectrum so the peak appears there. The signal x3[n] is a control that does not contain the target.

Modulation classification

Modulation classification entails identifying various schemes used to wirelessly transmit information, used in scenarios like cognitive radio. Using the same experimental setup in Fig. 2A, the input activation signal VX(1)(t) consists of a frame from a synthetically generated waveform across five different types of modulation: Gaussian frequency shift keying, continuous phase frequency shift keying, broadcast frequency modulation, double sideband amplitude modulation, and single sideband amplitude modulation as illustrated in Fig. 3D. The channel model includes fading due to Rician multipath, phase and frequency offsets due to clock offsets, timing drifts due to clock offsets, and AWGN such that the SNR is 30. This three-layer DNN experimentally achieved 85.0% single-shot modulation classification accuracy over 500 input activation frames compared to the digital single-shot accuracy of 89.2%.

Figure 3E illustrates the performance of a straightforward scheme to use multiple consecutive measurements to boost the accuracy using the single-shot experimental accuracy as the baseline. When MAFT-ONN makes multiple inferences, the final inference can be determined with a “majority vote” scheme, where the final inference is simply the class with the most inferences. The diagonal elements of the confusion matrix in Fig. 3D yield (C^=cC=c) , which is the conditional probability that the inference C^ will be correct given the input class C . The majority vote scheme assumes (i) that each measurement is statistically independent, (ii) the experimental single-shot accuracy represents the conditional probability for all incoming samples, (iii) the modulation class does not change in the duration of measurements, and (iv) only those five classes arrive at MAFT-ONN. With this scheme, the conditional probability of correct detection quickly approaches 100%, while reaching 95% with only 5 measurements.

We note that the chosen modulation classes are inherently more oriented toward frequency-domain encoding compared to modulation schemes like quadrature amplitude modulation, binary phase shift keying (BPSK), and quadrature phase shift keying (QPSK). When analyzing phase-modulated signals like BPSK in the frequency domain, the phase-encoded information is highly condensed around the carrier frequency. An input signal with information highly condensed in the frequency domain is problematic for MAFT-ONN due to the loss of computational density. As visualized in Fig. 1C, for a given input signal, increasing the frequency resolution of the weight kernel elements will increase the density of computations. Thus if the input signal is already too condensed in the frequency domain, creating weight signals with even denser frequency modes will incur an impractical amount of readout time.

This problem is alleviated by spreading the phase-encoded information into the frequency domain, enabling enough computational density to achieve the desired operation. As a proof of concept, we trained a digital twin of MAFT-ONN to distinguish between BPSK and QPSK with 90% accuracy with a single convolutional layer. In the digital twin, the phase information was spread into the frequency domain by appending an analog preprocessor before MAFT-ONN that applied an envelope detector and separately extracted the instantaneous frequency.

DISCUSSION

To our knowledge, the MAFT-ONN architecture is the first hardware accelerator that performs AI inference on raw communications signals without digitization or preprocessing. Our several experimental demonstrations show that MAFT-ONN is flexible enough to implement a powerful combination of fully analog frequency-domain LTI processing and AI inference to enable new high-performance signal processing capabilities with the benefits of the low power consumption, cost, size, and weight of optical systems. The power consumption of optical systems with different architectures but similar components has already been extensively studied (14, 23); thus, section SO analyzes the communication link gain of MAFT-ONN.

Computational throughput

The throughput T measures the number of multiply-and-accumulates (MACs) computed within a given time.

T=#MACslatency

The number of MACs computed in an FC layer with N input neurons and R output neurons is N·R MACs. A CONV layer can be interpreted as keeping the spurious frequencies in an FC layer; thus, the number of MACs in a CONV layer with N input neurons and N·R weight neurons is N2·R MACs. The time it takes to read out the output signal is the Gabor limit, which is 1min(Δf,f0) , where Δf is the smallest frequency spacing of the output signal and f0 is the lowest neuron frequency of the output signal. (Note that 2πf=ω , where ω is the angular frequency from the previous sections.)

Section SI derives the theoretical throughput, showing that the throughput for an FC layer asymptotically approaches the available bandwidth B for modulating signals: TFCB , and similarly for the throughput of a CONV layer: TCONVN·B.

Hence, for the MNIST classification experiment in Fig. 2A, the achieved throughput is T=3,851,600MACs120ns+1ms=3.85 giga-operations/s, where the 120 ns is the time-of-flight latency and the 1 ms is the readout latency due to the 1-kHz frequency spacing. The throughput of our experimental setup is limited by a bandwidth of 43 MHz, because the amplifier supports a minimum bandwidth of 2 MHz and the preceding balanced photodetector supports a maximum bandwidth of 45 MHz (see Methods for further component details). Given that the second balanced photodetector has a bandwidth of 150 MHz, the theoretical maximum throughput for the two CONV layers is 196·43MHz+10·150MHz10 giga-operations/second. The theoretical throughput can be achieved by exploiting the anti-aliasing characteristics near DC (section SI) and using wider frequency spacing (section SD).

For future scaling of MAFT-ONN, the bandwidth B limiting the throughput is not the RF bandwidth of the electrical components but the available optical bandwidth, for example, the 20 THz among the S, C, and L telecommunications bands. See section SM for an analysis of scalable MAFT-ONN architectures that take advantage of the full optical bandwidth and multiplexing.

Therefore, the combination of using the full optical bandwidth (on the order of terahertz) and spatial multiplexing (on the order of a hundred) (45) immediately yields a straightforward path to reaching peta-operation per second scale throughputs using MAFT-ONN with current technology. Thus, the MAFT-ONN is competitive with electronic counterparts like the Google TPUv3 that has a throughput greater than 400 tera-operations per second (46).

The throughput for a given bandwidth B is doubled by using complex-valued operations. The encoding scheme for computing complex-valued matrix products with MAFT-ONN is derived in section SL.

System-level latency

Figure 4A illustrates how state-of-the-art digital architectures require several steps to move data to and from the compute (where the compute elements are the red subblocks). This bottleneck is only being accentuated by Edholm’s law, which predicts that the amount of bandwidth and data rates for communications will continue to exponentially rise. Figure 4B illustrates an example of Edholm’s law for mobile communications, plotting the year each mobile network generation was introduced and its maximum bandwidth and data rate. We compare the theoretical system-level latency for various RF receiver architectures in Fig. 4C, where system-level latency is defined as the duration of time between the signal first arriving at the antenna to the time the decision-making system can extract the desired information from the output voltage signal.

Fig. 4. System-level latency analysis comparing the MAFT-ONN architecture to state-of-the-art digital architectures.

Fig. 4.

(A) Modeling various high-performance computing architectures in the context of receiving an RF signal x(t) , implementing a filter on it, and then sending it to a “decision system” that performs some operation based on the inferred qualities of the processed x(t) . The green highlights show that the analog information “flows” into and through MAFT-ONN, and thus the amount of information moving through the MAFT-ONN processor is limited by the Shannon capacity of the analog channels feeding it information (in the absence of nonideal factors like multipath fading and interference). (B) An example of Edholm’s law applied to mobile communications. This shows that wireless mobile bandwidth and data rates have been and will likely continue to exponentially grow. (C) Using the models in (A), the latency for processing x(t) with increasing bandwidth B is estimated for various high-performance processors. We assume that the digital processors are performing at peak specifications with 100% processor and memory utilization in ideal conditions. For simplicity, we assume that all FPGA components are running at 500 MHz. Because MAFT-ONN avoids the bottleneck of digital data movement, its overall latency improves by two orders of magnitude compared to traditional receivers.

In this scenario, we assume that the incoming signal x(t) has bandwidth B and that the computations for x(t) require frequency resolution Δf . For simplicity, we assume that the preselector circuit labeled “Rx” contributes negligible latency and that the ADC speed exactly matches the required Nyquist sampling rate for B . Together, B and Δf determine the number of points N=2BΔf required to digitally capture x(t) . Therefore, as B increases, the amount of digital data that must be moved to/from memory also increases, thereby raising the digital latency.

We used the models in Fig. 4A to estimate the theoretical latencies for the digital processors in Fig. 4C. We assumed that all processors and data lanes are operating: at their theoretical peak speeds, with 100% processor and memory utilization, with no processing or memory management overhead, with no data rate degradation from high clock speeds, with no throttling due to overheating, and limitless DRAM/HBM in which to store RF data. In practice, these processors may experience more than an order of magnitude slower latency than their theoretical optimal performance. The FGPAs will likely incur less overhead than the GPUs because the FGPAs trade off having less management overhead by storing less memory.

Per the scenario in Fig. 4A, after x(t) is processed into an output voltage signal, it reaches a “decision system” that performs the appropriate operation based on the output signal’s qualities. For the traditional receiver architectures, the output voltage signal is digital (this contribution to the digital latency is neglected). For MAFT, the output voltage signal is an analog waveform, so the decision system extracts the desired information from the waveform to perform its operation. MAFT’s decision system may or may not use an ADC, for example, a non-ADC decision system may consist of a bank of bandpass filters followed by an analog power selector circuit. Regardless of whether an ADC is used, the latency of MAFT’s decision system is fundamentally dominated by the Gabor limit.

As the zoom subplot in Fig. 4C illustrates, some digital processors theoretically have superior latency at low bandwidths, but even then are quickly overtaken by MAFT as the bandwidth increases. At 15 GHz, MAFT’s latency ranges from 7 to 11 times faster than digital architectures. As the input signal bandwidth increases, so does MAFT’s latency benefit. For example, when considering scenarios like 5G multiple-input multiple-output where there are massive volumes of RF data streaming into the receiver, the combined instantaneous bandwidth is much larger than that of a single antenna receiver. Figure 4C illustrates that at 1000 GHz, MAFT’s latency improvements range from 400 to 670 times faster than digital architectures, assuming the MAFT-ONN multiplexing architectures in section SM are used to reach 1000 GHz of bandwidth. The latency benefit from MAFT is physically attributed to the fast data movement to the compute. The green highlighted paths in Fig. 4A illustrate that the analog data “flows” into and through MAFT.

As demonstrated in the experiments portrayed in Fig. 3, the ability for MAFT to perform AI or linear processing on RF signals is agnostic to the type of modulation or coding scheme used for the input signal. Thus, the amount of data that can simultaneously be processed by MAFT is limited by the capacity of the wireless link feeding the input signal, given that the MAFT components can support that capacity. For the former, the capacity of a wireless link in a realistic scenario depends on several factors including attenuation from weather, multipath from structures, and wireless congestion. As shown in Fig. 4B, the peak data rate for which 5G is designed is 20 Gbps. Theoretically, given the 5-GHz bandwidth from Fig. 4B and a strong SNR of 25 dB, the theoretical Shannon capacity for a 5G link is ~40 Gbps. Studies show that 6G may yield 5 to 10 times (or higher) improvements over 5G data rate capacities (47, 48).

Therefore, for MAFT to process 5G and 6G signals, the optical components must achieve at least those capacities. While the theoretical Shannon capacity of such optical components is currently an unsolved problem (49, 50), there exist approximations such as the nonlinear Shannon limit (5052). The theoretical capacity of MAFT is best estimated by considering intensity-modulated direct detection (IMDD) links, which have been extensively studied and use the same optical components as MAFT. Data rates ranging from 100 to 200 Gbps per optical wavelength have been experimentally demonstrated for IMDD links (5355), showing that the optical components in the MAFT architecture fundamentally support such capacities. Hence, especially with the multiplexing schemes detailed in section SM, MAFT is comfortably poised to support 5G and 6G links at their capacities.

The time-of-flight latency per photoelectric multiplication of the MAFT-ONN experiment was measured to be 60 ns per layer, which matches the expected latency given the propagation through the DPMZM, fiber, and photodetector. The readout latency was 1 ms, determined by the Gabor limit. See section SP for a breakdown of our latency estimation models. Here, MAFT-ONN promises orders of magnitude faster latency compared to high-performance digital architectures while benefiting from the superior size, weight, power, and cost of photonic circuits. Because MAFT-ONN uses conventional components found in nanophotonic foundries, fabrication into a photonic integrated circuit is straightforward.

We have introduced and demonstrated our MAFT analog computing scheme and MAFT-ONN hardware accelerator for scalable, fully analog DNN acceleration that uses frequency-encoded neurons for convolutional and FC layers. The MAFT-ONN architecture yields much flexibility for running various types and sizes of DNNs without changing the hardware.

This architecture is also the first AI hardware accelerator that directly performs inference on raw RF signals without digitization or preprocessing, offering faster latency than FPGAs while simultaneously benefiting from the cost, size, weight, and power consumption of optics. The combination of frequency-domain LTI signal processing and trainable weight signals enables a flexible, high-performance computing paradigm. In addition, when using the full optical bandwidth and spatial multiplexing, the throughput of this processor is also competitive with other state-of-the-art AI hardware accelerators. Future work includes increasing the scale of MAFT-ONN using wavelength-division and spatial multiplexing.

METHODS

2 × 2 matrix multiplication experiment

For the example experiment in Fig. 1C that details a 2 × 2 matrix multiplication, VX(t) and VW(t) were generated by an arbitrary waveform generator (AWG) and then sent to DPMZMs without using voltage drivers. These DPMZMs perform SSB-SC modulation of the signals, without which the modulated signals would be dual-sideband and cancel each other out after the photoelectric multiplication.

To SSB-SC modulate a signal using a DPMZM, one copy of the signal is sent to one of the arms of the DPMZM while a 90 phase-shifted copy is sent to the other arm. Thus, let an underbar indicate an analytical Hilbert transform, yielding V_X(t)=Ha[VX(t)] . Then, Re[V_X(t)]=VX(t) is the original signal, and Im[V_X(t)] is the 90 phase-shifted copy. Although the 90° phase shift was generated using an AWG in this experiment, in practice, this phase shift can be achieved using commercial passive RF phase shifters.

Three-layer MNIST DNN

We experimentally demonstrated the three-layer DNNs using the MAFT-ONN scheme with the apparatus shown in Fig. 2A. An AWG generates VX(1)(t) , VW(1)(t) , and VW(2)(t) , all of which are modulated into the optical domain using DPMZMs directly from the AWG without using voltage drivers, as the 1.5-V peak output from the AWG was sufficient. Note that the accuracy can be improved by introducing drivers to increase the SNR by an order of magnitude because the 1.5-V peak AWG output corresponds to +16.5 dBm of power, whereas each DPMZM can safely handle a maximum power of +28 dBm. The photoelectric multiplication of VX(1)(t) and VW(1)(t) yielded Vout(1)(t) after the first layer. We then amplified Vout(1)(t) to reach the nonlinear regime of the DPMZM in the second layer, after which it became the input signal to the hidden layer. For convenience, we only used one submodulator of the DPMZM for this signal, thus modulating Vout(1)(t) in the dual-sideband suppressed carrier (DSB-SC) mode. For the hidden layer, we programmed VW(2)(t) as the next CONV kernel. The multiplication of the DSB-SC–modulated Vout(1)(t) and the SSB-SC–modulated VW(2)(t) results in a copy of Vout(2)(t) appearing further up in the spectrum, as can be seen in the plot of Y(2) in Fig. 2A. Last, the analog output of the second layer Vout(2)(t) was sampled digitally and Fourier transformed. See section SA for more details on the three-layer DNN experimental setup.

The input activations are downsampled 14 × 14 MNIST images that are represented by the frequency-encoded signal VX(1)(t) containing 196 frequencies spaced at 100 kHz. The input activation is convolved by a weight kernel VW(1)(t) containing 19,600 frequencies spaced at 1 kHz to yield the signal of the hidden layer, Vout(1)(t) . Note that VW(1)(t) was programmed as for an FC layer, but we chose to keep the “spurious” frequencies, thus effectively making it a CONV operation. Next, the hidden layer Vout(1)(t) , is multiplied by the second layer weight signal VW(2)(t) that contains 1000 frequencies spaced at 1 kHz to yield the output signal Vout(2)(t).

The signal VW(1)(t) was designed so that the frequency content of Vout(1)(t) would maximize use of the 43 MHz of bandwidth of the components. Thus, the frequency content of Vout(1)(t) is spread between 2 MHz (amplifier limit) and 45 MHz (photodetector limit). The frequency spacing of 1 kHz was used for VW(1)(t) and VW(2)(t) to maximize the number of MACs computed within that 43 MHz (thus increasing the expressivity of the DNN) while balancing out a reasonable readout time.

Physically, the minimum frequency spacing achievable is limited by the linewidth of the RF frequency modes, which may be widened from phase noise. Equipment-wise, it is limited by the RAM of the AWG for creating long time-series signals or the RAM of the data acquisition device. Application-wise, it is limited by the required readout speed.

We implemented the one-hot vector that represents the output MNIST values by randomly selecting a set of 10 adjacent frequencies among the frequencies of Vout(2)(t) to demonstrate the flexibility of our scheme. (See section SK for a performance analysis of using sparse FC layers for DNN training.) The 10 output neuron frequencies were randomly chosen to be 14.03 to 14.039 MHz, with 1 kHz spacing. The zoom of the plot of Vout(2)(t) in Fig. 2A shows the mapping of the neuron frequencies to the MNIST digits, where the digit is classified by the frequency mode with that largest magnitude. Hence, the final readout only considers the absolute value of the frequencies modes, making the one-hot vector insensitive to phase.

For the MNIST classification, we programmed real-valued positive and negative neuron values into both the input vectors and weight matrices. Negative neuron values are physically represented by a π phase shift in that particular frequency mode, allowing for analog matrix algebra with negative numbers.

An analytic model of the hardware was used to train the DNN offline, similar to previous ONN training schemes that account for the transfer function of the hardware (33, 56, 57). The offline training produced a set of weight matrices that were then encoded into the RF signals used for the experimental inferences. See sections SB, SC, and SE for details on the offline DNN characterization and training.

Three-layer modulation classification DNN

The same hardware setup used for the MNIST classification in Fig. 2A was also used to implement the three-layer DNN for modulation classification, so we reuse the mathematical notation. The input activation waveforms were generated using MATLAB (58). Each waveform had a random time delay applied to model receiving the signals at unknown times. The power of all the frames was normalized at baseband and then upconverted to 1 MHz as shown in Fig. 3C.

The first DNN layer was CONV with the DPMZM biases configured such that VW(1)(t) functioned as a trainable LTI filter. Following the first CONV layer was the DPMZM nonlinearity. Then, as portrayed in Fig. 2A, the second-layer weight signal VW(2)(t) was DSB-SC modulated and implemented another CONV layer. The LTI weight signal VW(1)(t) consisted of 50 frequency modes spaced at ~9 kHz and the second-layer weight signal VW(2)(t) consisted of 6143 frequency modes spaced at ~200 Hz. Both VW(1)(t) and VW(2)(t) were modeled as complex signals, so the number of trainable parameters was double the number of frequency modes (one for magnitude and phase).

For the one-hot vector readout of Vout(2)(t) , instead of choosing adjacent frequencies like in the MNIST classification, we chose frequencies that were evenly spread out across the spectrum as we found that this allowed the weights to train most effectively, and as with our MNIST inference, the absolute values of the frequencies were measured so that the frequency mode with the largest magnitude was considered the DNN inference.

Acknowledgments

S. Wilcox (MIT) aided in designing Fig. 1.

Funding: This work was supported by the Army Research Laboratory Electronic Warfare Branch (R.D.), the US Army (W911NF-18-2-0048, W911NF2120099, and W911NF-17-1-0527) (D.E. and R.D.), the US Air Force (FA9550-16-1-0391) (D.E. and R.D.), the MIT Lincoln Laboratory (PO\#7000442717) (D.E.), DARPA NaPSAC Program N66001-24-2-4002 (D.E.), Research Collaboration Agreements with Nippon Telegraph and Telephone (NTT) (D.E. and R.D.), and the National Science Foundation (NSF) (81350-Z3438201) (D.E. and R.D.).

Author contributions: D.E. and R.D. conceived the idea for the architecture. D.E. conceived the idea for programming neurons in time-frequency space and the method of using a single balanced photodetector for the photoelectric multiplication to compute the matrix-vector product in time-frequency space. R.D. conceived the MAFT frequency-encoding scheme that allows for single-shot photoelectric multiplication in the frequency domain by appropriately programming the inputs and weights and the method of using SSB-SC modulated signals for the MAFT-ONN hardware architecture. R.D. developed the theoretical performance metrics, conducted the hardware simulations, and planned and executed the experiments. Z.C. aided with planning and executing the hardware setup, characterization, debugging, and comparison with the theory. R.D. finalized and analyzed the experimental results. R.H. introduced the idea of using a discrete cosine transform to efficiently model the MZM nonlinearity, and R.D. conceived and created the offline physics-based DNN training algorithm. R.D. conceived and developed the LTI signal processing framework, the latency estimation model, modulation classification, and performed the various signal processing experiments. R.D. wrote and edited the manuscript with contributions from all authors. D.E. supervised the project.

Competing interests: R.D. and D.E. disclose that they are inventors on pending patent US application nos. 63/315,403 and 63/638,172 where MIT is the patent applicant. All other authors declare that they have no competing interests.

Data and materials availability: All data needed to evaluate the conclusions in the paper are present in the paper and/or the Supplementary Materials.

Supplementary Materials

This PDF file includes:

Supplementary Text

Figs. S1 to S14

Tables S1 to S7

References

sciadv.adt3558_sm.pdf (3.3MB, pdf)

REFERENCES AND NOTES

  • 1.Gao J., Wu D., Yin F., Kong Q., Xu L., Cui S., Metaloc: Learning to learn wireless localization. IEEE J. Sel. Areas in Commun. 41, 3831–3847 (2023). [Google Scholar]
  • 2.Jiao L., Shao Y., Sun L., Liu F., Yang S., Ma W., Li L., Liu X., Hou B., Zhang X., Shang R., Li Y., Wang S., Tang X., Guo Y., Advanced deep learning models for 6G: Overview, opportunities and challenges. IEEE Access 12, 133245–133314 (2024). [Google Scholar]
  • 3.de la Rosa J. M., AI-managed cognitive radio digitizers. IEEE Circ. Syst. Mag. 22, 10–39 (2022). [Google Scholar]
  • 4.G. C. Lee, A. Weiss, A. Lancho, J. Tang, Y. Bu, Y. Polyanskiy, G. W. Wornell. “Exploiting temporal structures of cyclostationary signals for data-driven single-channel source separation,” in 2022 IEEE 32nd International Workshop on Machine Learning for Signal Processing (MLSP) (IEEE, 2022), pp. 1–6. [Google Scholar]
  • 5.Youssef K., Bouchard L., Haigh K., Silovsky J., Thapa B., Valk C. V., Machine learning approach to rf transmitter identification. IEEE J. Radio Freq. Identif. 2, 197–205 (2018). [Google Scholar]
  • 6.G. Shen, J. Zhang, A. Marshall, L. Peng, X. Wang, “Radio frequency fingerprint identification for lora using spectrogram and CNN,” in IEEE INFOCOM 2021-IEEE Conference on Computer Communications (IEEE, 2021), pp. 1–10. [Google Scholar]
  • 7.Ren J., Jiang X., Regularized 2-D complex-log spectral analysis and subspace reliability analysis of micro-doppler signature for uav detection. Pattern Recognit. 69, 225–237 (2017). [Google Scholar]
  • 8.Peng H.-T., Lederman J. C., Xu L., de Lima T. F., Huang C., Shastri B. J., Rosenbluth D., Prucnal P. R., A photonics-inspired compact network: Toward real-time AI processing in communication systems. IEEE J. Sel. Top. Quantum Electron. 28, 1–17 (2022). [Google Scholar]
  • 9.Wong L. J., Clark W. H., Flowers B., Buehrer R. M., Headley W. C., Michaels A. J., An RFML ecosystem: Considerations for the application of deep learning to spectrum situational awareness. IEEE Open J. Commun. Soc. 2, 2243–2264 (2021). [Google Scholar]
  • 10.Piovesan N., López-Pérez D., De Domenico A., Geng X., Bao H., Debbah M., Machine learning and analytical power consumption models for 5g base stations. IEEE Commun Mag 60, 56–62 (2022). [Google Scholar]
  • 11.Shen Y., Harris N. C., Skirlo S., Prabhu M., Baehr-Jones T., Hochberg M., Sun X., Zhao S., Larochelle H., Englund D., Soljačić M., Deep learning with coherent nanophotonic circuits. Nat. Photon. 11, 441–446 (2017). [Google Scholar]
  • 12.Zhang H., Gu M., Jiang X. D., Thompson J., Cai H., Paesani S., Santagati R., Laing A., Zhang Y., Yung M. H., Shi Y. Z., Muhammad F. K., Lo G. Q., Luo X. S., Dong B., Kwong D. L., Kwek L. C., Liu A. Q., An optical neural chip for implementing complex-valued neural network. Nat. Commun. 12, 457 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.H. Bagherian, S. Skirlo, Y. Shen, H. Meng, V. Ceperic, M. Soljacic, On-chip optical convolutional neural networks, arXiv:1808.03303 [cs.ET] (2018).
  • 14.S. Bandyopadhyay, A. Sludds, S. Krastanov, R. Hamerly, N. Harris, D. Bunandar,M. Streshinsky, M. Hochberg, D. Englund, Single chip photonic deep neural network with accelerated training, arXiv:2208.01623 [cs.ET] (2022).
  • 15.Kovaios S., Roumpos I., Moralis-Pegios M., Giamougiannis G., Berciano M., Ferraro F., Bode D., Srinivasan S. A., Pantouvaki M., Pleros N., Tsakyridis A., Scaling photonic neural networks: A silicon photonic gemm leveraging a time-space multiplexed xbar. J. Lightwave Technol. 42, 7825–7833 (2024). [Google Scholar]
  • 16.Xu S., Wang J., Zou W., Optical convolutional neural network with WDM-based optical patching and microring weighting banks. IEEE Photonics Technol. Lett. 33, 89–92 (2021). [Google Scholar]
  • 17.Tait A. N., de Lima T. F., Zhou E., Wu A. X., Nahmias M. A., Shastri B. J., Prucnal P. R., Neuromorphic photonic networks using silicon photonic weight banks. Sci. Rep. 7, 7430 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Feldmann J., Youngblood N., Wright C. D., Bhaskaran H., Pernice W. H., All-optical spiking neurosynaptic networks with self-learning capabilities. Nature 569, 208–214 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Bangari V., Marquez B. A., Miller H., Tait A. N., Nahmias M. A., de Lima T. F., Peng H.-T., Prucnal P. R., Shastri B. J., Digital electronics and analog photonics for convolutional neural networks (DEAP-CNNs). IEEE J. Sel. Top. Quantum Electron. 26, 1–13 (2020). [Google Scholar]
  • 20.Xu X., Tan M., Corcoran B., Wu J., Boes A., Nguyen T. G., Chu S. T., Little B. E., Hicks D. G., Morandotti R., Mitchell A., Moss D. J., 11 tops photonic convolutional accelerator for optical neural networks. Nature 589, 44–51 (2021). [DOI] [PubMed] [Google Scholar]
  • 21.Feldmann J., Youngblood N., Karpov M., Gehring H., Li X., Stappers M., Le Gallo M., Fu X., Lukashchuk A., Raja A. S., Liu J., Wright C. D., Sebastian A., Kippenberg T. J., Pernice W. H. P., Bhaskaran H., Parallel convolutional processing using an integrated photonic tensor core. Nature 589, 52–58 (2021). [DOI] [PubMed] [Google Scholar]
  • 22.Sludds A., Bandyopadhyay S., Chen Z., Zhong Z., Cochrane J., Bernstein L., Bunandar D., Dixon P. B., Hamilton S. A., Streshinsky M., Novack A., Baehr-Jones T., Hochberg M., Ghobadi M., Hamerly R., Englund D., Delocalized photonic deep learning on the internet’s edge. Science 378, 270–276 (2022). [DOI] [PubMed] [Google Scholar]
  • 23.Hamerly R., Bernstein L., Sludds A., Soljačić M., Englund D., Large-scale optical neural networks based on photoelectric multiplication. Phys. Rev. X 9, 021032 (2019). [Google Scholar]
  • 24.Wang T., Ma S.-Y., Wright L. G., Onodera T., Richard B. C., McMahon P. L., An optical neural network using less than 1 photon per multiplication. Nat. Commun. 13, 123 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Farhat N. H., Psaltis D., Prata A., Paek E., Optical implementation of the hopfield model. Appl. Optics 24, 1469 (1985). [DOI] [PubMed] [Google Scholar]
  • 26.S. Kung, H. Liu, “An optical inner-product array processor for associative retrieval,” in Nonlinear Optics and Applications, Vol. 613 (International Society for Optics and Photonics, 1986), pp. 214–219. [Google Scholar]
  • 27.J. Ohta, M. Takahashi, Y. Nitta, S. Tai, K. Mitsunaga, K. Kjuma, “A new approach to a gaas/algaas optical neurochip with three layered structure,” in Proceeding of the IJCNN International Joint Conference on Neural Networks, Vol. 2 (IEEE, 1989), pp. 477–482. [Google Scholar]
  • 28.Zuo Y., Li B., Zhao Y., Jiang Y., Chen Y.-C., Chen P., Jo G.-B., Liu J., Du S., All-optical neural network with nonlinear activation functions. Optica 6, 1132–1137 (2019). [Google Scholar]
  • 29.L. Bernstein, A. Sludds, C. Panuski, S. Trajtenberg-Mills, R. Hamerly, D. Englund, Single-shot optical neural network, arXiv:2205.09103 [cs.ET] (2022). [DOI] [PMC free article] [PubMed]
  • 30.Khoram E., Chen A., Liu D., Ying L., Wang Q., Yuan M., Yu Z., Nanophotonic media for artificial neural inference. Photonics Res. 7, 823–827 (2019). [Google Scholar]
  • 31.Ashtiani F., Geers A. J., Aflatouni F., An on-chip photonic deep neural network for image classification. Nature 1, 501–506 (2022). [DOI] [PubMed] [Google Scholar]
  • 32.Chen Z., Sludds A., Davis R. III, Christen I., Bernstein L., Ateshian L., Heuser T., Heermeier N., Lott J. A., Reitzenstein S., Hamerly R., Englund D., Deep learning with coherent VCSEL neural networks. Nat. Photon. 17, 723–730 (2023). [Google Scholar]
  • 33.Zhu H. H., Zou J., Zhang H., Shi Y. Z., Luo S. B., Wang N., Cai H., Wan L. X., Wang B., Jiang X. D., Thompson J., Luo X. S., Zhou X. H., Xiao L. M., Huang W., Patrick L., Gu M., Kwek L. C., Liu A. Q., Space-efficient optical computing with an integrated chip diffractive neural network. Nat. Commun. 13, 1044 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Liu W., Fu T., Huang Y., Sun R., Yang S., Chen H., C-DONN: Compact diffractive optical neural network with deep learning regression. Opt. Express 31, 22127–22143 (2023). [DOI] [PubMed] [Google Scholar]
  • 35.Lin X., Rivenson Y., Yardimci N. T., Veli M., Luo Y., Jarrahi M., Ozcan A., All-optical machine learning using diffractive deep neural networks. Science 361, 1004–1008 (2018). [DOI] [PubMed] [Google Scholar]
  • 36.Luo X., Hu Y., Ou X., Li X., Lai J., Liu N., Cheng X., Pan A., Duan H., Metasurface-enabled on-chip multiplexed diffractive neural networks in the visible. Light Sci. Appl. 11, 158 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Tait A. N., De Lima T. F., Nahmias M. A., Miller H. B., Peng H.-T., Shastri B. J., Prucnal P. R., Silicon photonic modulator neuron. Phys. Rev. Appl. 11, 064043 (2019). [Google Scholar]
  • 38.George J. K., Mehrabian A., Amin R., Meng J., De Lima T. F., Tait A. N., Shastri B. J., El-Ghazawi T., Prucnal P. R., Sorger V. J., Neuromorphic photonics with electro-absorption modulators. Opt. Express 27, 5181–5191 (2019). [DOI] [PubMed] [Google Scholar]
  • 39.Williamson I. A., Hughes T. W., Minkov M., Bartlett B., Pai S., Fan S., Reprogrammable electro-optic nonlinear activation functions for optical neural networks. IEEE J. Sel. Top. Quantum Electron. 26, 1–12 (2019). [Google Scholar]
  • 40.Jha A., Huang C., Prucnal P. R., Reconfigurable all-optical nonlinear activation functions for neuromorphic photonics. Opt. Lett. 45, 4819–4822 (2020). [DOI] [PubMed] [Google Scholar]
  • 41.Huang C., Jha A., De Lima T. F., Tait A. N., Shastri B. J., Prucnal P. R., On-chip programmable nonlinear optical signal processor and its applications. IEEE J. Sel. Top. Quantum Electron. 27, 1–11 (2021). [Google Scholar]
  • 42.Crnjanski J., Krstić M., Totović A., Pleros N., Gvozdić D., Adaptive sigmoid-like and prelu activation functions for all-optical perceptron. Opt. Lett. 46, 2003–2006 (2021). [DOI] [PubMed] [Google Scholar]
  • 43.J. R. Basani, M. Heuck, D. R. Englund, S. Krastanov, All-photonic artificial neural network processor via non-linear optics, arXiv:2205.08608 [physics.optics] (2022).
  • 44.Wright L. G., Onodera T., Stein M. M., Wang T., Schachter D. T., Hu Z., McMahon P. L., Deep physical neural networks trained with backpropagation. Nature 601, 549–555 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 45.Streshinsky M., Novack A., Ding R., Liu Y., Lim A. E.-J., Lo P. G.-Q., Baehr-Jones T., Hochberg M., Silicon parallel single mode 48× 50 gb/s modulator and photodetector array. J. Lightwave Technol. 32, 4370–4377 (2014). [Google Scholar]
  • 46.S. Kumar, V. Bitorff, D. Chen, C. Chou, B. Hechtman, H. J. Lee, N. Kumar, P. Mattson, S. Wang, T. Wang, Y. Xu, Z. Zhou, Scale mlperf-0.6 models on google tpu-v3 pods, arXiv:1909.09756 [cs.LG] (2019).
  • 47.Chen S., Liang Y.-C., Sun S., Kang S., Cheng W., Peng M., Vision, requirements, and technology trend of 6g: How to tackle the challenges of system coverage, capacity, user data-rate and movement speed. IEEE Wirel Commun 27, 218–228 (2020). [Google Scholar]
  • 48.M. I. Maulana, M. Suryanegara, “Progress in 6G technology: A short review,” in 2023 6th International Conference of Computer and Informatics Engineering (IC2IE) (IEEE, 2023),pp. 36–41. [Google Scholar]
  • 49.Secondini M., Forestieri E., Scope and limitations of the nonlinear shannon limit. J. Lightwave Technol. 35, 893–902 (2017). [Google Scholar]
  • 50.Derevyanko S. A., Prilepsky J. E., Turitsyn S. K., Capacity estimates for optical transmission based on the non-linear fourier transform. Nat. Commun. 7, 12710 (2016). [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Essiambre R.-J., Kramer G., Winzer P. J., Foschini G. J., Goebel B., Capacity limits of optical fiber networks. J. Lightwave Technol. 28, 662–701 (2010). [Google Scholar]
  • 52.Ellis A. D., Zhao J., Cotter D., Approaching the non-linear shannon limit. J. Lightwave Technol. 28, 423–433 (2010). [Google Scholar]
  • 53.Caruso G., Cano I. N., Nesset D., Talli G., Gaudino R., Real-time 100 Gb/s PAM-4 for access links with up to 34 dB power budget. J. Lightwave Technol. 41, 3491–3497 (2023). [Google Scholar]
  • 54.Pang X., Salgals T., Louchet H., Che D., Gruen M., Matsui Y., Dippon T., Schatz R., Joharifar M., Krüger B., Pittala F., Fan Y., Udalcovs A., Zhang L., Yu X., Spolitis S., Bobrovs V., Popov S., Ozolins O., 200 Gb/s optical amplifier-free IM/DD transmissions using a directly modulated O-band DFB+R laser targeting LR applications. J. Lightwave Technol. 41, 3635–3641 (2023). [Google Scholar]
  • 55.Torres-Ferrera P., Rizzelli G., Wang H., Ferrero V., Gaudino R., Experimental demonstration of 100 Gbps/λ C-band direct-detection downstream PON using non-linear and CD compensation with 29 dB+ OPL over 0 km–100 km. J. Lightwave Technol. 40, 547–556 (2022). [Google Scholar]
  • 56.G. Mourgias-Alexandris, A. Tsakyridis, N. Passalis, M. Kirtas, A. Tefas, T. Rutirawut,F. Y. Gardes, N. Pleros, M. Moralis-Pegios, “25GMAC/sec/axon photonic neural networks with 7GHz bandwidth optics through channel response-aware training,” in 2021 European Conference on Optical Communication (ECOC) (IEEE, 2021), pp. 1–4. [Google Scholar]
  • 57.Moralis-Pegios M., Mourgias-Alexandris G., Tsakyridis A., Giamougiannis G., Totovic A., Dabos G., Passalis N., Kirtas M., Rutirawut T., Gardes F. Y., Tefas A., Pleros N., Neuromorphic silicon photonics and hardware-aware deep learning for high-speed inference. J. Lightwave Technol. 40, 3243–3254 (2022). [Google Scholar]
  • 58.MathWorks. Modulation classification with deep learning. https://www.mathworks.com/help/comm/ug/modulation-classification-with-deep-learning.html.
  • 59.Lesko D. M. B., Timmers H., Xing S., Kowligy A., Lind A. J., Diddams S. A., A six-octave optical frequency comb from a scalable few-cycle erbium fibre laser. Nat. Photonics 15, 281–286 (2021). [Google Scholar]
  • 60.Couny F., Benabid F., Roberts P., Light P., Raymer M., Generation and photonic guidance of multi-octave optical-frequency combs. Science 318, 1118–1121 (2007). [DOI] [PubMed] [Google Scholar]
  • 61.V. J. Urick, K. J. Williams, J. D. McKinney, Fundamentals of microwave photonics (John Wiley & Sons, 2015). [Google Scholar]
  • 62.Hochberg M., Baehr-Jones T., Wang G., Huang J., Sullivan P., Dalton L., Scherer A., Towards a millivolt optical modulator with nano-slot waveguides. Opt. Express 15, 8401–8410 (2007). [DOI] [PubMed] [Google Scholar]
  • 63.D. L. Bhargav, D. Achish, G. Yashwanth, N. P. Kalyan, K. Ashesh, “Prediction of signal drop due to rain at user cellular signal reception,” in Sentimental Analysis and Deep Learning (Springer, 2022), pp. 357–367. [Google Scholar]
  • 64.L. Chrostowski, M. Hochberg, Silicon Photonics Design: From Devices to Systems (Cambridge Univ. Press, 2015). [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Supplementary Text

Figs. S1 to S14

Tables S1 to S7

References

sciadv.adt3558_sm.pdf (3.3MB, pdf)

Articles from Science Advances are provided here courtesy of American Association for the Advancement of Science

RESOURCES