Skip to content
BY 4.0 license Open Access Published by De Gruyter September 5, 2022

Diffractive interconnects: all-optical permutation operation using diffractive networks

  • Deniz Mengu ORCID logo , Yifan Zhao , Anika Tabassum , Mona Jarrahi ORCID logo and Aydogan Ozcan ORCID logo EMAIL logo
From the journal Nanophotonics

Abstract

Permutation matrices form an important computational building block frequently used in various fields including, e.g., communications, information security, and data processing. Optical implementation of permutation operators with relatively large number of input–output interconnections based on power-efficient, fast, and compact platforms is highly desirable. Here, we present diffractive optical networks engineered through deep learning to all-optically perform permutation operations that can scale to hundreds of thousands of interconnections between an input and an output field-of-view using passive transmissive layers that are individually structured at the wavelength scale. Our findings indicate that the capacity of the diffractive optical network in approximating a given permutation operation increases proportional to the number of diffractive layers and trainable transmission elements in the system. Such deeper diffractive network designs can pose practical challenges in terms of physical alignment and output diffraction efficiency of the system. We addressed these challenges by designing misalignment tolerant diffractive designs that can all-optically perform arbitrarily selected permutation operations, and experimentally demonstrated, for the first time, a diffractive permutation network that operates at THz part of the spectrum. Diffractive permutation networks might find various applications in, e.g., security, image encryption, and data processing, along with telecommunications; especially with the carrier frequencies in wireless communications approaching THz-bands, the presented diffractive permutation networks can potentially serve as channel routing and interconnection panels in wireless networks.

1 Introduction

Permutation is one of the basic computational operations that has played a key role in numerous areas of engineering, e.g., computing [1], communications [2], encryption [3], data storage [4], remote sensing [5], and data processing [6]. Historically, electronic integrated circuits have been the established implementation medium for the permutation operation and other space-variant linear transformations, while the research on optical computing has been mainly focused on using the Fourier transform approximation of thin lenses covering various applications in space-invariant transformations, e.g., convolution/correlation. On the other hand, as photonic switching devices and optical waveguide technology have become the mainstream communication tools on high-end applications, e.g., fiber optic networks, supercomputers, and data centers, various approaches have been developed toward all-optical implementation of permutation operation and other space-variant transformations based on, e.g., Mach–Zehnder interferometers [7], optical switches [8], photonic crystals [9], holographically recorded optical elements [1012], off-axis lenslet arrays [13, 14], and arrays of periodic grating-microlens doublets [15]. The development of compact, low-power optical permutation and interconnection devices can have significant impact on next-generation communication systems, e.g., 6G networks [16, 17], as well as other applications such as optical data storage [18] and image encrypting cameras [1921].

With the widespread availability of high-end graphics processing units (GPU) and the massively growing amounts of data, the past decade has witnessed major advances in deep learning, dominating the field of digital information processing for various engineering applications including, e.g., image segmentation and classification [2225], natural language processing [26, 27], among others [28]. The statistical inference and function approximation capabilities of deep neural networks have also been exploited to produce state-of-the-art performance for computational inverse problems in many imaging and sensing applications including, e.g., microscopy [2935], quantitative phase imaging [3641], and others [4251]. Beyond these data processing tasks, deep learning can also provide task-specific solutions to challenging inverse optical design problems for numerous applications including nanophotonics [52, 53], metamaterials [54], imaging, and sensing [5560]. However, as the success and the applications of deep learning grow further, the electronic parallel computing platforms, e.g., GPUs, hosting deep neural networks and other machine learning algorithms have started to bring some limitations due to their power- and bandwidth-hungry operation. Moreover, the pace of the advances in computational capacity of the integrated circuits has fallen behind the exponential increase predicted by the Moore’s law [61]. These factors have fueled a tremendous amount of effort toward the development of optical machine learning schemes and other photonic computing devices that can partially reduce the computational burden on the electronics leading to power-efficient, massively parallel, high-speed machine learning systems. While most of the arising optical computing techniques rely on integrated photonic devices and systems compatible with the integrated waveguide technology [6268], an alternative option toward exploiting photons for machine learning and the related computing tasks is to use complex modulation media and free-space light propagation and diffraction, which is particularly suitable for visual computing applications where the information is already carried by the optical waves (e.g., of a scene or target object) in free space [69].

Motivated by these pressing needs, Diffractive Deep Neural Networks (D2NN) [70, 71] have emerged as an optical machine learning framework that utilizes deep learning to engineer light–matter interactions over a series of diffractive surfaces so that a desired statistical inference or deterministic computing task is realized all-optically as the light waves propagate through structured surfaces. According to this framework, the physical parameters determining the phase and/or amplitude of light over each independently controllable unit, i.e., the “diffractive neuron,” are updated through the stochastic gradient descent and error-backpropagation algorithms based on a loss function tailored specifically for a given task. The weights of the connections between the diffractive neurons/features on successive layers, on the other hand, are dictated by the light diffraction in free space. Once the deep learning-based training, which is a one-time effort, is completed using a computer, the resulting transmissive/reflective diffractive layers are fabricated using, e.g., lithography or 3D printing, to physically form the diffractive network that completes a given inference or computational task at the speed of light using entirely passive modulation surfaces, offering a task-specific, power-efficient, and fast optical machine learning platform.

Based on the D2NN framework, here we demonstrate diffractive optical network designs that were trained to all-optically perform a given permutation operation between the optical intensities at the input and output fields-of-view, capable of handling hundreds of thousands of interconnections with diffraction limited resolution. We quantified the success of the presented diffractive optical networks in approximating a given, randomly selected permutation operation as a function of the number of diffractive neurons and transmissive layers used in the diffractive network design. We also laid the foundations toward practical implementations of diffractive permutation networks by investigating the impact of various physical error sources, e.g., lateral and axial misalignments and unwanted in-plane layer rotations, on the quality/accuracy of the optically realized interconnection weights and the permutation operation. Moreover, we showed that the diffractive optical permutation networks can be trained to be resilient against possible misalignments as well as imperfections in the diffractive layer fabrication and assembly. Finally, we present the first proof-of-concept experimental demonstration of diffractive permutation networks by all-optically achieving a permutation matrix of size 25 × 25, effectively realizing 625 interconnections based on 3D-printed diffractive layers operating at the THz part of the spectrum.

The presented diffractive optical permutation networks can readily find applications in THz-band communication systems serving as communication channel patch panels; furthermore, the underlying methods and design principles can be broadly extended to operate at other parts of the electromagnetic spectrum, including the visible and IR wavelengths, by scaling each diffractive feature size proportional to the wavelength of light [72] and can be used for image encryption in security cameras [21] and optical data storage systems, among other applications [7376].

2 Results

Figure 1 illustrates the presented free-space permutation interconnect concept designed around diffractive optical networks using the D2NN framework. As shown in Figure 1, the presented permutation interconnect scheme does not use any standard optical components such as lenses, and instead relies on a series of passive, phase-only diffractive surfaces. Due to the passive nature of these layers, the diffractive optical network shown in Figure 1 does not consume any power except for the illumination light, providing a power-efficient permutation operation in a compact footprint of ∼600λ along the longitudinal axis, which could be further squeezed as needed. The 5-layer diffractive optical permutation network design shown in Figure 1 was trained through supervised learning to all-optically realize a desired permutation operation, P , between the light intensity signals at the input and output FOVs, each with N i = N o = 400 (20 × 20) individual pixels of size 2λ × 2λ. Stated differently, this permutation operation controls in total of N i N o = 0.16 million optical intensity connections.

Figure 1: 
The schematic of a 5-layer diffractive permutation network, all-optically realizing 0.16 million interconnects between an input and output field-of-view. The presented diffractive permutation network was trained to optically realize an arbitrarily selected permutation operation between the light intensities over N

i
 = 400 = 20 × 20 input and N

o
 = 400 = 20 × 20 output pixels, establishing N

i

N

o
 = 0.16 million desired interconnections based on 5 phase-only diffractive layers, each containing 40K(200 × 200) diffractive neurons/features.
Figure 1:

The schematic of a 5-layer diffractive permutation network, all-optically realizing 0.16 million interconnects between an input and output field-of-view. The presented diffractive permutation network was trained to optically realize an arbitrarily selected permutation operation between the light intensities over N i = 400 = 20 × 20 input and N o = 400 = 20 × 20 output pixels, establishing N i N o = 0.16 million desired interconnections based on 5 phase-only diffractive layers, each containing 40K(200 × 200) diffractive neurons/features.

The supervised nature of the training process of the diffractive permutation network necessitates the use of a set of input–output signal pairs (examples that satisfy P ) to compute a penalty term and the associated gradient-based updates with respect to the physical parameters of each diffractive neuron at every iteration. We set the optical signal of interest at the input and the output of the diffractive permutation scheme to be the light intensity, and as a result, the deep learning-based evolution of the presented diffractive permutation network shown in Figure 1 was driven based on the mean-squared error (MSE) (see Methods Section) between the ground-truth and the all-optically synthesized output intensity patterns at a given iteration. Since this loss function acts only on the light intensity, the diffractive optical network can enjoy an output phase-freedom in synthesizing the corresponding transformed optical intensity patterns within the output field-of-view. The light intensity, I, is related to the complex-valued field, U, through a nonlinear operation, I = U 2 . If a pair of input–output complex fields exists for a given diffractive network, i.e., {U in, U out} and { U i n , U out }, then the input field U i n = α U i n + β U i n will create the output field U out = α U out + β U out at the output plane. In terms of the associated intensities, however, this direct linear extension does not hold since α U out 2 + β U out 2 U out 2 , making it challenging (in terms of the data generalization capability) to design diffractive optical networks for achieving a general purpose intensity-to-intensity transformation such as a permutation operation. To overcome this generalization challenge, we trained our diffractive permutation networks using ∼4.7 million randomly generated input/output intensity patterns that satisfy the desired P , instead of a standard benchmark image dataset (see the Methods).

After the training phase, we blindly tested each diffractive permutation network with test inputs that were never used during the training. Figure 2 illustrates 6 different randomly generated blind testing inputs along with the corresponding all-optically permuted output light intensities. In the first two randomly generated input patterns shown in Figure 2A, there is light coming out of all the input pixels/apertures at different levels of intensity. In the next two test input patterns shown in Figure 2A, on the other hand, nearly half of the input apertures have nonzero light intensity and finally, the last two test inputs contain only 10 and 11 pixels/apertures with light propagating toward the 1st layer of the diffractive permutation network. When tested on 20K randomly generated blind testing input intensity patterns with different sparsity levels, the 5-layer diffractive permutation network shown in Figure 1 achieves 18.61 dB peak-signal-to-noise ratio (PSNR), very well matching the ideal output response. For this randomly generated 20K testing data, Figure 2B also shows the distribution of PSNR as a function of the number of input pixels with nonzero light intensity, which reveals that the diffractive permutation network can permute relatively sparser inputs with a higher output image quality, achieving a PSNR of 25.82 dB.

Figure 2: 
Input–output intensity pairs generated by the diffractive permutation network shown in Figure 1. (A) The diffractive permutation network shown in Figure 1 was tested on two different datasets. The first blind testing dataset contains 20K randomly generated inputs. 6 examples from this randomly created testing data are shown here for demonstrating input–output intensity pairs with low, moderate and high signal sparsity levels. Beyond successfully permuting randomly generated intensity patterns, the performance of the diffractive permutation network was also quantified using permuted EMNIST images. None of these test samples were used in the training phase. (B) Output intensity image PSNR with respect to the ground truth intensity patterns as a function of the input signal sparsity in randomly generated test dataset. (C) Same as (B), except for EMNIST test images.
Figure 2:

Input–output intensity pairs generated by the diffractive permutation network shown in Figure 1. (A) The diffractive permutation network shown in Figure 1 was tested on two different datasets. The first blind testing dataset contains 20K randomly generated inputs. 6 examples from this randomly created testing data are shown here for demonstrating input–output intensity pairs with low, moderate and high signal sparsity levels. Beyond successfully permuting randomly generated intensity patterns, the performance of the diffractive permutation network was also quantified using permuted EMNIST images. None of these test samples were used in the training phase. (B) Output intensity image PSNR with respect to the ground truth intensity patterns as a function of the input signal sparsity in randomly generated test dataset. (C) Same as (B), except for EMNIST test images.

In addition to randomly generated blind testing inputs, we further tested the diffractive permutation network shown in Figure 1 on 18.75K EMNIST images; note that this diffractive network was trained only using randomly generated input/output intensity patterns that satisfy P and the EMNIST images constitute not only blind testing set but also a significant deviation from the statistical distribution of the training images. The input field-of-view contains the permuted EMNIST images ( P −1) and the diffractive network inverts that permutation by all-optically performing P to recover the original images at the output plane (see Figure 2). The performance of the diffractive permutation network was quantified based on both PSNR and Structural Similarity Index Measure (SSIM). With N L = 40K diffractive neurons on each layer, the 5-layer diffractive permutation network shown in Figure 1 provides 19.18 dB and 0.85 for PSNR and SSIM metrics, respectively, demonstrating the generalization capability of the diffractive network to new types of input image data never seen during the training phase.

2.1 Impact of the number of diffractive layers and features

Next, we investigate the performance of diffractive permutation networks as a function of the number of diffractive neurons trained within the system. Toward this goal, in addition to the 5-layer design shown in Figure 1 that has in total of N = 200K trainable diffractive features, we trained diffractive permutation networks consisting of 4, 3, and 2 modulation surfaces. The physical design parameters such as the size/width of the diffractive surfaces, layer-to-layer distances, and the extent of the input and output fields-of-view were kept identical to the ones in the 5-layer network design. In other words, these new diffractive networks are designed and trained exactly in the same way as the previous 5-layer network except they contain fewer diffractive layers. Figure 3A and B provides a quantitative comparison between these 4 diffractive permutation networks. While Figure 3A illustrates the mean PSNR and SSIM values achieved by each diffractive network for recovering EMNIST images, Figure 3B demonstrates the mean-squared-error (MSE) between the desired permutation operation and its optically realized version ( P D 2 N N ) as a function of the number of diffractive layers utilized in these designs (see Supplementary Figure S1 and the Methods Section on the estimation of P D 2 N N ). According to the permutation operator error shown in Figure 3B, the performance improvement of the system increases drastically with the additional diffractive layers up to the 4-layer design that represents a critical point in the sense that the inclusion of a 5th diffractive surface brings a relatively small improvement. The reason behind this behavior is the fact that the number of diffractive features, N, in the 4-layer diffractive permutation network matches the space-bandwidth product set by our input and output FOVs, i.e., N i N o = 400 × 400 = 160K. In other words, Figure 3B reveals that when the number of phase-only diffractive modulation units N matches or exceeds N i N o , the diffractive optical network can achieve a given linear transformation between the input and output intensities with a very low error, i.e., P D 2 N N P ; for example, the MSE between P D 2 N N and P in the case of a 4-layer design was found to be 6.63 × 10−5. For N < N i N o , the error between P D 2 N N and P increases accordingly, as shown in Figure 3B.

Figure 3: 
The impact of the number of diffractive layers on the approximation accuracy of D2NN for a given intensity permutation operation. (A) The average SSIM and PSNR values achieved by the diffractive permutation network designs based on L = 2, L = 3, L = 4 and L = 5 diffractive layers, containing 200 × 200, i.e., 40K, phase-only diffractive neurons/features per layer for the task of optically recovering permuted EMNIST images. (B) The transformation error between the desired intensity permutation (P) and its optically realized counterpart (




P




D


2


N
N




${\boldsymbol{P}}_{{\boldsymbol{D}}^{2}\boldsymbol{N}\boldsymbol{N}}$



) for the diffractive networks with L = 2, L = 3, L = 4 and L = 5 diffractive layers. The transformation error decreases as a function of the number of layers in the diffractive network architecture. The L = 4-layer diffractive permutation network design represents a critical point as it matches the space-bandwidth product requirement of the desired permutation operation, i.e., N = N

i

N

o
 = 4 × 40K = 160K, and further increasing the number of layers to L = 5 brings only a minor improvement. (C) Examples of EMNIST test images demonstrating the performance of the diffractive permutation networks as a function of L.
Figure 3:

The impact of the number of diffractive layers on the approximation accuracy of D2NN for a given intensity permutation operation. (A) The average SSIM and PSNR values achieved by the diffractive permutation network designs based on L = 2, L = 3, L = 4 and L = 5 diffractive layers, containing 200 × 200, i.e., 40K, phase-only diffractive neurons/features per layer for the task of optically recovering permuted EMNIST images. (B) The transformation error between the desired intensity permutation (P) and its optically realized counterpart ( P D 2 N N ) for the diffractive networks with L = 2, L = 3, L = 4 and L = 5 diffractive layers. The transformation error decreases as a function of the number of layers in the diffractive network architecture. The L = 4-layer diffractive permutation network design represents a critical point as it matches the space-bandwidth product requirement of the desired permutation operation, i.e., N = N i N o = 4 × 40K = 160K, and further increasing the number of layers to L = 5 brings only a minor improvement. (C) Examples of EMNIST test images demonstrating the performance of the diffractive permutation networks as a function of L.

The benefit of having NN i N o is further revealed in the increased generalization capability of the diffractive network as shown in Figure 3A. Since the EMNIST images were not used during the training, they represent completely new types of input intensity patterns for the presented diffractive optical networks. The SSIM (PSNR) values achieved by the 4-layer diffractive network is found as 0.75 (16.41 dB) for the optical recovery of the permuted EMNIST images. These numbers are significantly higher compared to the performance of the 3-layer and 2-layer diffractive designs that can attain SSIM (PSNR) values of 0.46 (12.91 dB) and 0.30 (12.08 dB) for the same task; furthermore, the 5-layer diffractive network design shown in Figure 1 outperforms the others by achieving 0.85 (19.18 dB) for the same performance metrics. The visual comparison of the input–output intensity patterns depicted in Figure 3C further supports this conclusion, where the noise due to the crosstalk between interconnection channels decreases proportional to the number of diffractive layers in the system.

2.2 Vaccination of diffractive permutation networks

With sufficiently large number of phase-only diffractive neurons/features, the diffractive networks can optically realize permutation operations with, e.g., 0.16 million channels between the input and output pixels as shown in Figure 3. In fact, the number of interconnects that can be optically implemented through diffractive networks can go far beyond 0.16 million, given that the size/width of the diffractive surfaces and the number of diffractive layers can be increased further depending on the fabrication technology and the optomechanical constraints of the system. In addition, as the number of diffractive layers increases in a diffractive network architecture, their forward model can better generalize to new, unseen data as shown in Figure 3A.

On the other hand, deeper diffractive optical network designs are more susceptible to misalignments that are caused by the limitations of the optomechanical assembly and/or the fabrication technology that is utilized. It was shown that diffractive optical networks trained for statistical inference tasks, e.g., all-optical object classification, can be vaccinated against misalignments and other physical error sources, when the factors creating these nonideal conditions were incorporated into the training forward model, which was termed as vaccinated-D2NNs or v-D2NNs [77]. Specifically, v-D2NN expands on the original D2NN framework by modeling possible error sources as random variables and integrating them as part of the training model so that the deep learning-based evolution of the diffractive surfaces is guided toward solutions that are resilient to nonideal physical conditions and/or fabrication errors. Toward practical applications of diffractive permutation networks, we quantified the impact of optomechanical errors and applied the v-D2NN framework to devise robust solutions that can achieve a given interconnect operation despite fabrication tolerances.

In our numerical study depicted in Figure 4, we considered 4 different misalignment components representing the 3D misalignment vector of the lth diffractive layer, D x l , D y l , D z l and their in-plane rotation around the optical axis denoted as D θ l . Each of these 4 misalignment components were defined as independent, uniformly distributed random variables, D * l U ( Δ * , Δ * ) , with Δ* defined as a function of a common auxiliary parameter, v. The lateral misalignments parameters, Δ x and Δ y , determining the range of D x l and D y l , respectively, were set to be 0.67λv, i.e., D x l U ( 0.67 λ v , 0.67 λ v ) and D y l U ( 0.67 λ v , 0.67 λ v ) , where λ denotes the wavelength of the illumination light. Similarly, Δ z and Δ θ were defined as 24λv and 4°v. For instance, if we take v = 0.5, this means each diffractive layer can independently/randomly shift in both x and y axes within a range of (−0.335λ, 0.335λ). In addition, their location over the z direction and their in-plane orientation can randomly change within the ranges of (−12λ, 12λ) and (−2°, 2°), respectively (see the Methods Section for more details).

Figure 4: 
The sensitivity of the diffractive permutation networks against various levels of physical misalignments (A) SSIM values achieved by 5-layer diffractive permutation networks with and without vaccination. (B) Transformation errors between the desired 100 × 100 permutation operation (P) and its optically synthesized counterpart (




P




D


2


N
N




${\boldsymbol{P}}_{{\boldsymbol{D}}^{2}\boldsymbol{N}\boldsymbol{N}}$



) at different levels of misalignments denoted by v
test. (C) The layers of a nonvaccinated diffractive permutation network, i.e., v
tr = 0, along with the examples of EMNIST test images recovered optically through the diffractive permutation operation. (D) Same as (C), except for a vaccinated diffractive permutation network based on v
tr = 0.25.
Figure 4:

The sensitivity of the diffractive permutation networks against various levels of physical misalignments (A) SSIM values achieved by 5-layer diffractive permutation networks with and without vaccination. (B) Transformation errors between the desired 100 × 100 permutation operation (P) and its optically synthesized counterpart ( P D 2 N N ) at different levels of misalignments denoted by v test. (C) The layers of a nonvaccinated diffractive permutation network, i.e., v tr = 0, along with the examples of EMNIST test images recovered optically through the diffractive permutation operation. (D) Same as (C), except for a vaccinated diffractive permutation network based on v tr = 0.25.

To better highlight the impact of these misalignments and demonstrate the efficacy of the v-D2NN framework, we trained a new nonvaccinated, i.e., v tr = 0, diffractive permutation network that can all-optically realize a given permutation matrix, P , representing 10K intensity interconnections between 100 input and 100 output pixels of size 4λ × 4λ. The error-free training model of this diffractive network with v tr = 0 implicitly assumes that when the resulting diffractive network is fabricated, the system conditions will exactly match the ideal settings regarding the 3D locations of the layers and their in-plane orientations. With an architecture identical to the one shown in Figure 1, containing N = 200K ≫ N i N o diffractive neurons, this diffractive network can all-optically approximate the permutation matrix, P , with an MSE of 1.45 × 10−6 in the absence of any misalignment errors, i.e., v test = 0 (see the green curve in Figure 4B). However, when there is some discrepancy between the training and testing conditions, i.e., v test > 0, the optically implemented forward transformation, P D 2 N N , starts to deviate from the desired operation P . For instance, at v test = 0.125, the transformation error, P D 2 N N P , can be computed as 3.1 × 10−3. This negative impact of the physical misalignments on the performance of a nonvaccinated diffractive network can also be seen in Figure 4A (green curve), which demonstrates the SSIM values achieved by this diffractive network for recovering permuted EMNIST images under different levels of misalignments. The high-quality of the image recovery (see Figure 4C) at v test = 0 quantified with an SSIM of 0.99 deteriorates under the presence of misalignments, highlighted by the SSIM value falling to 0.49 and 0.30 at v test = 0.125 and v test = 0.25, respectively.

Unlike the nonvaccinated design, the vaccinated diffractive permutation networks can maintain their approximation capacity and accuracy under erroneous testing conditions as shown in Figure 4A and B. For instance, the SSIM value of 0.49 attained by the nonvaccinated diffractive network for the misalignment uncertainty set by v test = 0.125 increases to 0.88 in the case of a diffractive permutation network trained with v tr = 0.25 (red curve in Figure 4A). The difference between the image recovery performances of the vaccinated and the nonvaccinated diffractive network designs increases further as the misalignment levels increase during the blind testing. While the nonvaccinated diffractive network can only achieve SSIM values of 0.3 and 0.19 at v test = 0.25 and v test = 0.375, respectively, the output images synthesized by the vaccinated design (v tr = 0.25) reveals SSIM values of 0.8 at v test = 0.25 and 0.64 at v test = 0.375 (see Figure 4D). A similar conclusion can also be drawn from Figure 4B, demonstrating the MSE values between the desired permutation matrix, P , and its optically realized counterpart, P D 2 N N . The transformation errors, P D 2 N N P , of the vaccinated diffractive network (v tr = 0.25) at v test = 0.125 and at v test = 0.25 were computed as 5.15 × 10−4 and 1.2 × 10−3, respectively, which are 5–10 times smaller compared to the MSE values provided by the nonvaccinated diffractive design at the same misalignment levels. Supplementary Figure S2 further illustrates the error maps between P and P D 2 N N realized by the nonvaccinated and vaccinated diffractive permutation networks at different misalignment levels.

The compromise for this misalignment robustness comes in the form of a reduction in the peak performance. While the nonvaccinated diffractive network can solely focus on realizing the given permutation operation with the highest quality and approximation accuracy, the vaccinated diffractive network designs partially allocate their degrees-of-freedom to building up resilience against physical misalignments. For example, while the peak SSIM achieved by the nonvaccinated diffractive network is 0.99, it is 0.88 for the diffractive permutation network vaccinated with v tr = 0.25. The key difference, on the other hand, is that the better performance of the nonvaccinated diffractive network is sensitive to the physical implementation errors, while the vaccinated diffractive permutation networks can realize the desired input–output interconnects over a larger range of fabrication errors or tolerances. A comparison between the diffractive layer patterns of the nonvaccinated and vaccinated diffractive permutation networks shown in Figure 4C and D, respectively, also reveals that the vaccination strategy results in smoother light modulation patterns; in other words, the material thickness values over the neighboring diffractive neurons partially lose their independence and become correlated, causing a reduction in the number of independent degrees-of-freedom in the system.

2.3 Experimental demonstration of a diffractive permutation network

To experimentally demonstrate the success of the presented diffractive permutation interconnects, we designed a 3-layer diffractive permutation network achieving the desired (randomly generated) intensity shuffling operation with N i = N o = 5 × 5, optically synthesizing 625 connections between the input and output FOVs; this network was designed to operate at 0.4 THz, corresponding to ∼0.75 mm in wavelength. During the training, the forward model of this diffractive permutation network was vaccinated with v tr = 0.5 against the 4 error sources as detailed in Section 2.2 including the 3D location of each diffractive layer and the in-plane rotation angle around the optical axis. In addition to these misalignment components, we also vaccinated this diffractive network model against unwanted material thickness variations that could arise due to the limited lateral and axial resolution of our 3D printer (see the Methods Section for details). To compensate for the reduction in the degrees-of-freedom due to the vaccination scheme, the number of phase-only diffractive features in the permutation network was selected to be N L = 10K diffractive neurons per layer. Therefore, each diffractive layer shown in Figure 5A contains 100 × 100 phase-only diffractive neurons of size ∼0.67λ × 0.67λ. Compared to the diffractive surfaces shown in Figures 1 4, the layers of our experimental system were set to be 2-times smaller in both the x and y directions to keep the layer-to-layer distances smaller while maintaining the level of optical connectivity between the successive diffractive surfaces (see Figure 5B). Figure 5C and D illustrates the 3D printed diffractive permutation network and the schematic of our experimental setup (see the Methods Section for details).

Figure 5: 
Experimental demonstration of a diffractive permutation network. (A) The material thickness profiles of the diffractive surfaces of the fabricated diffractive permutation network. (B) The schematic of the experimental architecture illustrating the forward optical model of the diffractive permutation network. (C) 3D printed diffractive permutation network operating at THz part of the spectrum. (D) The schematic of our experimental system.
Figure 5:

Experimental demonstration of a diffractive permutation network. (A) The material thickness profiles of the diffractive surfaces of the fabricated diffractive permutation network. (B) The schematic of the experimental architecture illustrating the forward optical model of the diffractive permutation network. (C) 3D printed diffractive permutation network operating at THz part of the spectrum. (D) The schematic of our experimental system.

Figure 6A illustrates the targeted 25 × 25 permutation matrix ( P ) that is randomly generated and the numerically predicted P D 2 N N along with the absolute difference map between these two matrices. According to the numerical forward model of the trained diffractive network shown in Figure 5, the transformation error between the P and P D 2 N N , i.e., P D 2 N N P is equal to 5.99 × 10−4 under error-free conditions, i.e., v test = 0. Furthermore, the forward model of the trained diffractive permutation network shown in Figure 5 provides 17.87 dB PSNR on average for the test letters “U,” “C,” “L,” and “A,” as depicted in Figure 6B. A visual comparison between the numerically predicted and the experimentally measured output images of these 4 input letters (which were never seen by the network before) demonstrates the accuracy of the forward training and testing models as well as the success of the presented diffractive permutation network design. Interestingly, the PSNR of the experimentally measured images was observed to be higher, 19.54 dB, compared to the numerically predicted value, 17.87 dB. Our numerical study reported in Figure 4 suggests that this can be explained based on the vaccination range used during the training and the amount physical error in the system testing. For instance, the SSIM value achieved by the vaccinated diffractive network trained with v tr = 0.5 (yellow curve) at relatively lower physical misalignment levels, e.g., v test = 0.125, is higher compared to its performance under the ideal conditions, i.e., v test = 0.0, as depicted in Figure 4A.

Figure 6: 
Experimental results. (A) (left) The desired 25 × 25 permutation matrix, P, (middle) the optically realized permutation operation predicted by the numerical forward model, 




P




D


2


N
N




${\boldsymbol{P}}_{{\boldsymbol{D}}^{2}\boldsymbol{N}\boldsymbol{N}}$



, and (right) the absolute error map between the two matrices. (B) Comparison between the numerically predicted and the experimentally measured output images for the task of recovering intensity patterns describing the letters “U,” “C,” “L,” and “A”.
Figure 6:

Experimental results. (A) (left) The desired 25 × 25 permutation matrix, P, (middle) the optically realized permutation operation predicted by the numerical forward model, P D 2 N N , and (right) the absolute error map between the two matrices. (B) Comparison between the numerically predicted and the experimentally measured output images for the task of recovering intensity patterns describing the letters “U,” “C,” “L,” and “A”.

3 Discussion

Beyond optomechanical error sources and fabrication tolerances, another factor that might potentially hinder the utilization of diffractive permutation networks in practical applications is the output diffraction efficiency. For instance, the diffraction efficiency of the 5-layer network shown in Figure 1 is ∼0.004% which might be very low for some applications. On the other hand, this can be significantly increased by using an additional loss term, penalizing the poor diffraction efficiency of the network (see the Methods Section for details). Supplementary Figure S3 demonstrates 5-layer diffractive networks that are designed to optically realize 0.16 million interconnections between the input and output FOVs with increased diffraction efficiencies, and compares their performance in terms of SSIM values achieved. The training of these diffractive network models is based on a loss function in the form of a linear combination of two different penalty terms, L = L + γ L e , where L L e is the diffraction efficiency related penalty term promoting efficient solutions (see the Methods Section). As a general trend, the diffraction efficiency of the underlying diffractive network model increases as a function of the weight (γ) of the efficiency penalty term in the loss function. However, since the number of diffractive neurons, hence, the degrees-of-freedom in these diffractive network models is very close to N i N o , the diffraction efficiency either does not improve beyond a certain value or the evolution of the diffractive layers starts to solely focus on the efficiency instead of the desired permutation operation resulting in low performance designs. This unstable behavior can be observed specifically when 0.235 < γ < 0.24. On the other hand, as in the case of vaccinated diffractive network models, if the diffractive network architecture contains NN i N o diffractive neurons, then this instability vanishes, providing significant improvements in the output diffraction efficiency without sacrificing the performance of the all-optical permutation operation. For instance, the 3D printed diffractive permutation network depicted in Figure 6 was trained based on L with γ = 0.15 and it provides 2.45% output diffraction efficiency, despite the fact that 89.37% of the incident power at the input plane is lost due to the absorption of the 3D printing material. With weakly absorbing transparent materials used as part of the diffractive network fabrication, a significantly larger output efficiency can be achieved.

Also note that, although we solely focused on diffractive network designs composed of dielectric optical modulation surfaces, in principle, some of these layers can be replaced with metasurfaces/metamaterials. While the use of can provide some additional degrees of freedom, including, for example, the engineering of dispersion, there are some challenges to overcome in realizing metamaterial-based diffractive networks. In the presence of fabrication errors and imperfections, the scattered light fields might deviate from the predictions of the numerical forward model. These nonideal waves generated by a metasurface would then excite unwanted diffraction modes over the subsequent layers generating an “avalanche” within the diffractive volume, accumulating substantial field errors, especially for deeper network designs with L ≥ 2. In addition, the physical models of phase and/or amplitude modulation of meta-atoms are, in general, valid for waves covering a relatively small numerical aperture (NA). As a result, a high NA diffractive network design that utilizes the entire bandwidth of the propagation medium (NA = 1, for air) would be challenging, as the modulation response of the meta-units might deviate from their ideal small angle responses, introducing errors to the forward model. Although it is possible to restrict a diffractive network design to work with a lower NA, it would increase the overall footprint of the system and reduce the space-bandwidth product that can be processed by the diffractive network.

These challenges, in general, are negligible for dielectric diffractive networks composed of λ/2 features on a substrate, as also highlighted by the close match between the numerically predicted images and their experimentally measured counterparts shown in Figure 6 and our former work [21, 70, 74, 76, 77]. In our optical forward model, the diffractive layers are assumed to be thin modulation surfaces, i.e., there is only a single scattering event converting an incident wave field to an outgoing one after each diffractive layer. Practically, though, there are additional scattered fields that are ignored in our model, especially when there is a substantial material thickness variation between adjacent pixels. However, in our designs, we do not observe a sharp material thickness transition between neighboring pixels. This is mainly due to the nature of our training process. Specifically, the presented diffractive networks are trained through error-backpropagation, which computes the variable updates by taking the gradient of the loss function with respect to the material thickness over each diffractive unit. In such a process, it is highly unlikely that the gradients of two adjacent pixels (λ/2 apart from each other) deviate significantly from each other, which effectively causes a smoothening effect on the diffractive surface height profiles as they are being optimized through deep learning. This smoothening behavior is even more pronounced in the vaccinated diffractive network designs due to the random lateral translation of the diffractive layers as part of the training forward model. Therefore, the impact of side scattering or field shadowing due to height discontinuities across a given diffractive layer design is negligible. In addition, the back-reflected waves can also be ignored, as these are, in general, weak processes unless they are specifically enhanced using, e.g., metamaterials or other special structures. Therefore, the optical forward model of dielectric diffractive networks can be accurately represented within the scalar diffraction theory without needing vectorial modeling of light fields or considering weaker multi-reflections. Finally, the evanescent waves and the vectorial fields associated with them can be entirely ignored since each successive diffractive layer is axially positioned >λ away from the previous layer.

In summary, we showed that the diffractive networks can optically implement intensity permutation operations between their input and output apertures based on phase-only light modulation surfaces with NN i N o diffractive neurons. Due to the nonlinear nature of the intensity operation, it is crucial to use training input intensity patterns with different levels of sparsity to prevent any type of data-specific overfitting during the training phase. Diffractive permutation networks with N > N i N o demonstrate increased generalization capability, synthesizing more accurate outputs with P D 2 N N P 0 . By using N > N i N o , one can also design misalignment and fabrication error insensitive, power-efficient diffractive permutation networks, which could play a major role in practical applications, e.g., 6G wireless networks, computational cameras, etc. Although this study demonstrated diffractive optical networks realizing permutation operations with 0.16 million interconnects, with N i = N o = 20 × 20, these systems are highly scalable to even larger N i , N o combinations depending on the availability of training computer hardware. Since the training of a diffractive optical network is a one-time effort, one can use a computing platform with a significantly larger random-access memory (RAM) to design much bigger diffractive networks. Alternatively, the forward training model of a diffractive network can also be distributed among multiple GPUs for parallel computing with increased memory capacity paving the way to significantly larger permutation operations to be implemented all-optically. Finally, the incorporation of dynamic spatial light modulators to replace some of the diffractive layers in a given design can be used to reconfigure, on demand, the all-optically performed diffractive transformation.

4 Methods

4.1 Experimental setup

According to the schematic diagram of our experimental setup shown in Figure 5D, the THz wave incident on the input FOV of the diffractive network was generated using a horn antenna attached to the source WR2.2 modulator amplifier/multiplier chain (AMC) from Virginia Diode Inc. (VDI). A 10 dBm RF input signal at 11.111 GHz (f RF1) at the input of the AMC was multiplied 36 times to generate a continuous-wave (CW) radiation at 0.4 THz, corresponding to ∼0.75 mm in wavelength. The output of the AMC was modulated with 1 kHz square wave to resolve low-noise output data through lock-in detection. Since we did not use any collimating optics in our setup, the distance between the input plane of the 3D-printed diffractive optical network and the exit aperture of the horn antenna was set to be ∼60 cm approximating a uniform plane wave over the 40λ × 40λ input FOV. At the output plane of the diffractive optical network, the diffracted THz light was collected using a single-pixel Mixer/AMC from Virginia Diode Inc. (VDI). During the measurements, the detector received a 10 dBm sinusoidal signal at 11.083 GHz serving as a local oscillator for mixing, and the down-converted signal was at 1 GHz. The 40λ × 40λ output FOV was scanned by placing the single-pixel detector on an XY stage that was built by combining two linear motorized stages (Thorlabs NRT100). At each scan location, the down-converted signal coming from the single-pixel detector was fed to low-noise amplifiers (Mini-Circuits ZRL-1150-LN+) with a gain of 80 dBm and a 1 GHz (+/−10 MHz) bandpass filter (KL Electronics 3C40-1000/T10-O/O) that erases the noise components coming from unwanted frequency bands. Following the amplification and filtering, the measured signal passed through a tunable attenuator (HP 8495B) and a low-noise power detector (Mini-Circuits ZX47-60). Finally, the output voltage value was generated by a lock-in amplifier (Stanford Research SR830). The modulation signal was used as the reference signal for the lock-in amplifier and accordingly, we performed a calibration to convert the lock-in amplifier readings at each scan location to linear scale. During our experiments, the scanning step size at the output plane was set to be ∼λ in x and y directions. The smallest pixel of the experimentally targeted permutation grid, i.e., the desired resolution of the diffractive permutation operation was taken as 8λ × 8λ during the training, corresponding to 5 × 5 discrete input and output signals. Therefore, the output signal measured for each input object was integrated over a region of 8λ × 8λ per pixel, resulting in the measured images shown in Figure 6B.

A 3D printer, Objet30 Pro, from Stratasys Ltd., was used to fabricate the layers of the diffractive permutation network shown in Figure 5C as well as the layer holders. The active modulation area of our 3D printed diffractive layers was 5 × 5 cm (∼66.66λ × ∼66.66λ) containing 100 × 100, i.e., 10K, diffractive neurons. These modulation surfaces were printed as insets surrounded by a uniform slab of printing material with a thickness of 2.5 mm and the total size of each printed layer including these uniform regions was 6.2 × 6.2 cm. Following the 3D printing, these additional surrounding regions were coated with aluminum to block the propagation of the light over these areas minimizing the contamination of the output signal with unwanted scattered light.

4.2 Training forward model of diffractive permutation networks

4.2.1 Optical forward model

The material thickness, h, was selected as the physical parameter controlling the complex-valued transmittance values of the diffractive layers of our design. Based on the complex-valued refractive index of the diffractive material, τ = n + , the corresponding transmission coefficient of a diffractive neuron located on the lth layer at a coordinate of (x q , y q , z l ) is defined as,

(1) t x q , y q , z l = exp 2 π κ h x q , y q , z l λ × exp j 2 π n n m h x q , y q , z l λ

where n m = 1 denotes the refractive index of the propagation medium (air) between the layers. The real and imaginary parts of the 3D printing material were measured experimentally using a THz spectroscopy system, and they were revealed as n = 1.7227 and κ = 0.031 at 0.4 THz.

The optical forward model of the presented diffractive networks relies on the Rayleigh–Sommerfeld theory of scalar diffraction to represent the propagation of light waves between the successive layers. According to this diffraction formulation, the free space can be interpreted as a linear, shift-invariant operator with the impulse response,

(2) w x , y , z = z r 2 1 2 π r + n j λ exp j 2 π n r λ

where r = x 2 + y 2 + z 2 . Based on Eq. (2), qth diffractive neuron on the lth layer, at (x q , y q , z l ), can be interpreted as the source of a secondary wave generating the field at (x, y, z) in the form of,

(3) w q l x , y , z = z z l r q l 2 1 2 π r q l + n j λ exp j 2 π n r q l λ .

The parameter r q l in Eq. (3) is expressed as x x q 2 + y y q 2 + z z l 2 . When each diffractive neuron on layer l generates the field described by Eq. (3), the light field incident on the pth diffractive neuron on the (l + 1)th layer at (x p , y p , z l+1) is the linear superposition of the all the secondary waves generated by the previous layer l, i.e., q A q l w q l x p , y p , z l + 1 , where A q l is the complex amplitude of the wave field right after the qth neuron of the lth layer. This field is modulated by the multiplicative complex-valued transmittance of the diffractive unit at (x p , y p , z l+1), creating the modulated field t x p , y p , z l + 1 q A q l w q l x p , y p , z l + 1 . Based on this new modulated field, a new secondary wave,

(4) u p l + 1 x , y , z = w p l + 1 x , y , z t x p , y p , z l + 1 q A q l w q l × x p , y p , z l + 1 ,

is generated. The outlined successive modulation and secondary wave generation processes occur until the waves propagating through the diffractive network reach to the output plane. Although, the forward optical model described by Eqs. (1)(4) is given over a continuous 3D coordinate system, during our deep learning-based training of the presented diffractive permutation networks, all the wave fields and the modulation surfaces were represented based on their discrete counterparts with a spatial sampling rate of ∼0.67λ on both x and y axes, that is also equal to the size of a diffractive neuron.

4.2.2 Physical architecture of the diffractive permutation networks and training loss functions

The size of the output and input FOVs of the presented diffractive permutation networks were both set to be 40λ × 40λ, defining a unit magnification optical permutation operation. Note that the unit magnification is not a necessary condition for the success of the forward operation of diffractive optical interconnects but rather a design choice. Without loss of generality, the output FOV can be defined centered around the origin, 0,0 , i.e., − 20λ < x, y < 20λ. The dimensions of the diffractive layers was taken as 133.3 λ × 133.3λ for the diffractive permutation networks presented in Figures 1 4 and Supplementary Figure S3, and in all these diffractive network architectures, the layer-to-layer distances were taken as 120λ. The axial distance between the 1st diffractive layer and the input FOV was set to be 53.3λ that is also equal to the axial distance from the last diffractive layer to the output plane, preserving the symmetry of the system on the longitudinal axis. In the case of our experimentally validated diffractive design (Figure 5), on the other hand, the active modulation surface of the fabricated diffractive layers extends 66.7λ on both x and y directions. Accordingly, the layer-to-layer distances were taken as 60λ while the remaining distances were kept equal to 53.3λ.

During the deep learning-based training of all of these diffractive permutation networks, the wave fields and the propagation functions depicted in Eqs. (2)(4) were sampled at a rate of ∼0.67λ that is also equal to the size of the smallest diffractive units on the modulation surfaces constituting the presented diffractive networks. At this spatial sampling rate, the input and output intensity patterns were represented as 2D discrete vectors of size 60 × 60 denoted by I i n m , n and I out m , n , respectively, with m = 1, 2, 3, …, 60 and n = 1, 2, 3, …, 60. The underlying complex-valued wave fields can be written as U i n m , n = I i n m , n e j i n m , n and U out m , n = I out m , n e j out m , n . In our forward model, we assumed that the input light has constant phase front, i.e., i n m , n is taken as an arbitrary constant within the input field-of-view. In alternative implementations, without loss of generality, the diffractive permutation network can be trained with any arbitrary function of i n m , n , achieving the same output accuracy levels P D 2 N N P 0 using NN i N o .

While the light fields, the diffractive layers, and the impulse response of the free space were all sampled at a rate of ∼0.67λ, the spatial grid/pixel size of a given desired permutation operation was set to be larger. Specifically, the permutation pixel size was taken as 2λ × 2λ for the diffractive networks shown in Figures 1 3. On the other hand, the input and output pixel size was chosen as 4λ × 4λ for the vaccinated and nonvaccinated diffractive permutation networks shown in Figure 4, and finally, the pixel size was set to be 8λ × 8λ for the fabricated diffractive permutation network model depicted in Figure 5.

To train the presented diffractive permutation networks, a structural loss function, L , in the form of MSE was used.

(5) L = 1 S s = 1 S P I i n [ s ] σ I out [ s ] 2 ,

In Eq. (5), I i n s and I out s denote the lexicographically ordered vectorized counterparts of the input intensity pattern, i.e., v e c I i n q , p , and the output intensity pattern, i.e., v e c I out q , p , and P represents the desired permutation matrix to be performed all-optically. As depicted in Eq. (5), the output intensity pattern I out[s] or I out q , p was scaled by a constant σ that was calculated at each training iteration as,

(6) σ = 1 S s = 1 S P I i n [ s ] I out [ s ] 1 S s = 1 S I out [ s ] 2 .

Note that the presented diffractive permutation networks preserve the relative intensity levels. Stated differently, our training forward model aims to keep the intensity levels over the output and input pixels the same up to a single multiplicative constant, σ.

To improve the diffraction efficiency of diffractive permutation networks, we defined another loss function, L , that is a linear combination of two penalty terms, L = L + γ L e , where L corresponds to the structural loss defined in Eq. (5). L e is the penalty term that promotes higher diffraction efficiency at the output of diffractive networks, and it was defined as, L e = e η , where,

(7) η = s = 1 S I out [ s ] s = 1 S I i n [ s ] × 100

The diffractive permutation networks presented in Figures 1 4 were trained based on L with γ = 0; however, the experimentally demonstrated diffractive permutation network model was trained with γ = 0.15, resulting in an output diffraction efficiency of 2.45%, which includes a material absorption loss of Supplementary Figure S3 further demonstrates the diffraction efficiency and the SSIM values provided by various diffractive permutation network models trained with different γ values.

The supervised deep learning-based training of the presented diffractive permutation networks evaluates the loss function L for a batch of randomly generated input patterns, computes the mean gradient, and updates the learnable, auxiliary variables, h a, that determine the material thickness over each diffractive neuron, h, through the following relation,

(8) h h a = sin h a + 1 2 ( h m h b ) + h b

where h m and h b denote the maximum modulation thickness and the base material thickness, respectively. For all the diffractive permutation networks presented in Figures 1 4 and Supplementary Figure S3, h m was taken as 2λ. In the design of the 3D-printed diffractive permutation network, however, h m was set to be 1.66λ to restrict the material thickness contrast between the neighboring diffractive features. The value of h b was taken as 0.66λ for all the presented designs including the fabricated diffractive network.

4.2.3 Computation of P D 2 N N , optical transformation errors and performance quality metrics

For a given diffractive permutation network design trained to optically implement a permutation matrix P of size N i × N o , there are two different ways to compute the permutation operation predicted by its numerical forward model. The first way is to propagate N different randomly generated independent inputs with NN i N o and solve a linear system of equations for revealing the entries of P D 2 N N . Alternatively, each input pixel at the input FOV can be turned on sequentially and the output intensity pattern synthesized by the diffractive optical permutation network as a response to each pixel provides one unique column of P D 2 N N . These two procedures, in general, result in two different P D 2 N N matrices that closely resemble each other. We opted to use the latter procedure due to its simplicity, which turn on each input pixel one at a time and records the corresponding output intensity pattern, which, after vectorization, represents a column of P D 2 N N . Following the calculation of P D 2 N N predicted by the forward model of a trained diffractive permutation network, it was scaled with a multiplicative constant, σ P , to account for the optical losses:

(9) σ P = 1 N i N o n i N i n o N o P D 2 N N n i , n o P n i , n o 1 N i N o n i N i n o N o P D 2 N N n i , n o 2 .

The all-optical transformation error, P P D 2 N N 2 , can be computed based on,

(10) P P D 2 N N 2 = 1 N i N o n i N i n o N o σ P P D 2 N N n i , n o P n i , n o 2

Denoting the lexicographically ordered vectorized version of a 2D input intensity pattern with I in[s], the ground truth output intensity can be found by P I in[s]. The PSNR between this ground-truth vector and the output vector synthesized by the forward optical operation of a given, trained diffractive network, I out[s], can be calculated as,

(11) P S N R = 20 l o g 10 1 s P I i n [ s ] σ I out [ s ] 2 ,

where σ is the multiplicative constant defined in Eq. (6). The SSIM values were calculated based on the built-in function in TensorFlow, i.e., tf.image.ssim, where the two inputs were 2D versions of P I in[s] and I out[s], representing the ground-truth image and the permuted, all-optical output signal, respectively. All the parameters of tf.image.ssim were taken equal to default values, except that the size of the Gaussian filter was set to be 5 × 5, instead of 11 × 11, and the width of the Gaussian filter was set to be 0.75.

4.2.4 Vaccination of diffractive permutation networks

v-D2NN framework aims to design diffractive optical networks that are resilient against physical error sources, e.g., misalignments, by modeling these factors as random variables and incorporating them into the forward training model. In the training forward model of the vaccinated diffractive networks shown in Figure 4, 4 physical error components were modeled representing the misalignment of each diffractive layer with respect to their ideal location and orientation/angle. The first 3 components represent the statistical variations in the location of each diffractive layer in 3D space. Let the ideal location of a diffractive layer, l, be denoted by the vector X l = x l , y l , z l , then at each training iteration i, v-D2NN framework perturbs X l with a random displacement vector, D l , i = D x l , i , D y l , i , D z l , i . The components of this 3D displacement vector were defined as uniformly distributed, independent random variables, i.e.,

(12a) D x l , i U ( Δ x , Δ x )

(12b) D y l , i U ( Δ y , Δ y )

(12c) D z l , i U ( Δ y , Δ y )

During the training, for each batch of input images, the 3D displacement vector D l , i is updated and accordingly, the location of the layer l is set to be X l , i = X l + D l , i , building up robustness to physical misalignments.

Beyond the displacement of diffractive layers, the physical forward model of a diffractive network is also susceptible to variations in the orientation of the diffractive layers. Ideally, one should include all 3 rotational components, yaw, pitch, and roll; however, in this study, we only considered the yaw component since in our experimental systems, the pitch and the roll can be controlled with a high precision. The random angle representing the rotation of a diffractive layer l around the optical axis was defined as D θ l , i U ( Δ θ , Δ θ ) . With 3 shift components depicted in Eq. (12) and the statistical yaw variation modeled through D θ l , i , the vaccinated diffractive networks shown in Figure 4 were trained to build resilience against these 4 misalignment components. The values of Δ x , Δ y , Δ z , and Δ θ determining the misalignment tolerance range were defined as a function a common variable v, i.e., Δ x = Δ y = 0.67λv, Δ z = 24λv, and Δ θ = 4°.

For the design of the experimentally validated diffractive permutation network, on top of these 4 optomechanical error components (with v = 0.5), we also modeled fabrication errors in the form of statistical variations of the material thickness over each diffractive neuron (h). Hence, at a given iteration, i, the material thickness values over each diffractive unit h h a , defined in Eq. (8), was perturbed through h i h a = h h a + D h i , where D h i U ( 0.025 h m , 0.025 h m ) . Stated differently, the fabricated diffractive layers shown in Figure 5 were designed to be resilient against physical errors on the material thickness values over the diffractive neurons within a range [−0.0415λ, 0.0415λ].

4.2.5 Training details

The deep learning-based training of the diffractive permutation networks was implemented using Python (v3.6.5) and TensorFlow (v1.15.0, Google Inc.). The backpropagation updates were calculated using the Adam optimizer [78], and its parameters were taken as the default values in TensorFlow and kept identical in each model. The learning rates of the diffractive optical networks were set to be 0.001. The training batch size was taken as 75 during the deep learning-based training of the presented diffractive permutation networks. For the training of the diffractive permutation networks, we generated ∼4.7 million random intensity patterns, providing us 93,750 iterations/error-backpropagation updates per epoch. The training of a 5-layer diffractive permutation network with 40K diffractive neurons per layer for 5 epochs using this randomly created training dataset takes approximately 4 days using a computer with a GeForce GTX 1080 Ti Graphical Processing Unit (GPU, Nvidia Inc.) and Intel® Core™ i7-8700 Central Processing Unit (CPU, Intel Inc.) with 64 GB of RAM, running Windows 10 operating system (Microsoft).


Corresponding author: Aydogan Ozcan, Electrical and Computer Engineering Department, University of California, Los Angeles, CA, 90095, USA; Bioengineering Department, University of California, Los Angeles, CA, 90095, USA; California NanoSystems Institute, University of California, Los Angeles, CA, 90095, USA; and Department of Surgery, David Geffen School of Medicine, University of California, Los Angeles, CA, 90095, USA, E-mail:

Award Identifier / Grant number: FA9550-21-1-0324

Acknowledgments

The Ozcan Research Group at UCLA acknowledges the support of AFOSR, Materials with Extreme Properties Program funding (FA9550-21-1-0324).

  1. Author contributions: All the authors have accepted responsibility for the entire content of this submitted manuscript and approved submission.

  2. Research funding: This work was supported by Air Force Office of Scientific Research (grant number FA9550-21-1-0324).

  3. Conflict of interest statement: The authors declare no conflicts of interest regarding this article.

References

[1] N. Ishikawa, S. Sugiura, and L. Hanzo, “50 Years of permutation, spatial and index modulation: from classic RF to visible light communications and data storage,” IEEE Commun. Surv. Tutor., vol. 20, pp. 1905–1938, 2018. https://doi.org/10.1109/comst.2018.2815642.Search in Google Scholar

[2] S. Ishimura and K. Kikuchi, “Multi-dimensional permutation-modulation format for coherent optical communications,” Opt. Express, vol. 23, p. 15587, 2015. https://doi.org/10.1364/oe.23.015587.Search in Google Scholar

[3] H. Huang, X. He, Y. Xiang, W. Wen, and Y. Zhang, “A compression-diffusion-permutation strategy for securing image,” Signal Process., vol. 150, pp. 183–190, 2018. https://doi.org/10.1016/j.sigpro.2018.04.014.Search in Google Scholar

[4] A. Jiang, R. Mateescu, M. Schwartz, and J. Bruck, “Rank modulation for flash memories,” IEEE Trans. Inform. Theory, vol. 55, pp. 2659–2673, 2009. https://doi.org/10.1109/tit.2009.2018336.Search in Google Scholar

[5] X. Huang, G. Ye, H. Chai, and O. Xie, “Compression and encryption for remote sensing image using chaotic system,” Secur. Commun. Netw., vol. 8, pp. 3659–3666, 2015. https://doi.org/10.1002/sec.1289.Search in Google Scholar

[6] T. K. Moon and W. C. Sterling, Mathematical Methods and Algorithms for Signal Processing, Upper Saddle River, NJ, Prentice Hall, 2000.Search in Google Scholar

[7] J. Carolan, C. Harrold, C. Sparrow, et al.., “Universal linear optics,” Science, vol. 349, pp. 711–716, 2015. https://doi.org/10.1126/science.aab3642.Search in Google Scholar PubMed

[8] R. A. Spanke and V. E. Benes, “N-stage planar optical permutation network,” Appl. Opt., vol. 26, p. 1226, 1987. https://doi.org/10.1364/ao.26.001226.Search in Google Scholar PubMed

[9] M. Djavid, M. H. T. Dastjerdi, M. R. Philip, et al.., “Photonic crystal-based permutation switch for optical networks,” Photonic Netw. Commun., vol. 35, pp. 90–96, 2018. https://doi.org/10.1007/s11107-017-0719-7.Search in Google Scholar

[10] H. Kobolla, F. Sauer, and R. Volkel, Holographic Tandem Arrays, G. M. Morris, Ed., Bellingham, Washington, USA, SPIE, 1989, p. 146.10.1117/12.961677Search in Google Scholar

[11] B. Robertson, E. J. Restall, M. R. Taghizadeh, and A. C. Walker, “Space-variant holographic optical elements in dichromated gelatin,” Appl. Opt., vol. 30, p. 2368, 1991. https://doi.org/10.1364/ao.30.002368.Search in Google Scholar PubMed

[12] H. Kobolla, J. T. Sheridan, E. Gluch, et al.., “Holographic 2D mixed polarization deflection elements,” J. Mod. Opt., vol. 40, pp. 613–624, 1993. https://doi.org/10.1080/09500349314550671.Search in Google Scholar

[13] M. C. Hutley, P. Savander, and M. Schrader, “The use of microlenses for making spatially variant optical interconnections,” Pure Appl. Opt., vol. 1, pp. 337–346, 1992. https://doi.org/10.1088/0963-9659/1/6/006.Search in Google Scholar

[14] J. Jahns and W. Däschner, “Optical cyclic shifter using diffractive lenslet arrays,” Opt. Commun., vol. 79, pp. 407–410, 1990. https://doi.org/10.1016/0030-4018(90)90472-6.Search in Google Scholar

[15] F. Sauer, J. Jahns, C. R. Nijander, A. Y. Feldblum, and W. P. Townsend, “Refractive-diffractive micro-optics for permutation interconnects,” Opt. Eng., vol. 33, p. 1550, 1994. https://doi.org/10.1117/12.166528.Search in Google Scholar

[16] A. Tarable, F. Malandrino, L. Dossi, R. Nebuloni, G. Virone, and A. Nordio, “Meta-surface optimization in 6G sub-THz communications,” in 2010 IEEE International Conference on Communications Workshops (ICC Workshops), IEEE, 2020, pp. 1–6.10.1109/ICCWorkshops49005.2020.9145267Search in Google Scholar

[17] X. Xu, Q. Chen, X. Mu, Y. Liu, and H. Jiang, “Graph-embedded multi-agent learning for smart reconfigurable THz MIMO-NOMA networks,” IEEE J. Sel. Areas Commun., vol. 40, pp. 259–275, 2022. https://doi.org/10.1109/jsac.2021.3126079.Search in Google Scholar

[18] B. M. King and M. A. Neifeld, “Sparse modulation coding for increased capacity in volume holographic storage,” Appl. Opt., vol. 39, p. 6681, 2000. https://doi.org/10.1364/ao.39.006681.Search in Google Scholar PubMed

[19] R. Enayatifar, A. H. Abdullah, I. F. Isnin, A. Altameem, and M. Lee, “Image encryption using a synchronous permutation-diffusion technique,” Opt. Lasers. Eng., vol. 90, pp. 146–154, 2017. https://doi.org/10.1016/j.optlaseng.2016.10.006.Search in Google Scholar

[20] V. Patidar, N. K. Pareek, G. Purohit, and K. K. Sud, “A robust and secure chaotic standard map based pseudorandom permutation-substitution scheme for image encryption,” Opt. Commun., vol. 284, pp. 4331–4339, 2011. https://doi.org/10.1016/j.optcom.2011.05.028.Search in Google Scholar

[21] B. Bai, Y. Luo, T. Gan, et al.., “To image, or not to image: class-specific diffractive cameras with all-optical erasure of undesired objects,” eLight, vol. 2, no. 1, p. 14, 2022. https://doi.org/10.1186/s43593-022-00021-3.Search in Google Scholar

[22] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” Commun. ACM, vol. 60, pp. 84–90, 2017. https://doi.org/10.1145/3065386.Search in Google Scholar

[23] Y. LeCun, L. Bottou, Y. Bengio, and P. Ha, “Gradient-based learning applied to document recognition,” Proc. IEEE, vol. 86, pp. 2278–2374, 1998. https://doi.org/10.1109/5.726791.Search in Google Scholar

[24] L. C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “DeepLab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected CRFs,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 40, pp. 834–848, 2018. https://doi.org/10.1109/tpami.2017.2699184.Search in Google Scholar PubMed

[25] O. Ronneberger, P. Fischer, and T. Brox, “U-net: convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, vol. 9351, N. Navab, J. Hornegger, W. M. Wells, and A. F. Frangi, Eds., Springer International Publishing, 2015, pp. 234–241.10.1007/978-3-319-24574-4_28Search in Google Scholar

[26] R. Collobert and J. Weston, “A unified architecture for natural language processing: deep neural networks with multitask learning,” in Proceedings of the 25th International Conference on Machine Learning – ICML’08, ACM Press, 2008, pp. 160–167.10.1145/1390156.1390177Search in Google Scholar

[27] Y. Goldberg, “Neural network methods for natural language processing,” Synth. Lect. Hum. Lang. Technol., vol. 10, pp. 1–309, 2017. https://doi.org/10.2200/s00762ed1v01y201703hlt037.Search in Google Scholar

[28] I. Goodfellow, Y. Bengio, and A. Courville, Deep Learning, Cambridge, Massachusetts, The MIT Press, 2016.Search in Google Scholar

[29] Y. Rivenson, Z. Göröcs, H. Günaydin, Y. Zhang, H. Wang, and A. Ozcan, “Deep learning microscopy,” Optica, vol. 4, p. 1437, 2017. https://doi.org/10.1364/optica.4.001437.Search in Google Scholar

[30] Y. Rivenson, H. C. Koydemir, H. Wang, et al.., “Deep learning enhanced mobile-phone microscopy,” ACS Photonics, vol. 5, pp. 2354–2364, 2018. https://doi.org/10.1021/acsphotonics.8b00146.Search in Google Scholar

[31] K. de Haan, Z. S. Ballard, Y. Rivenson, Y. Wu, and A. Ozcan, “Resolution enhancement in scanning electron microscopy using deep learning,” Sci. Rep., vol. 9, p. 12050, 2019. https://doi.org/10.1038/s41598-019-48444-2.Search in Google Scholar PubMed PubMed Central

[32] E. Nehme, L. E. Weiss, T. Michaeli, and Y. Shechtman, “Deep-STORM: super-resolution single-molecule microscopy by deep learning,” Optica, vol. 5, p. 458, 2018. https://doi.org/10.1364/optica.5.000458.Search in Google Scholar

[33] E. Nehme, D. Freedman, R. Gordon, et al.., “DeepSTORM3D: dense 3D localization microscopy and PSF design by deep learning,” Nat. Methods, vol. 17, pp. 734–740, 2020. https://doi.org/10.1038/s41592-020-0853-5.Search in Google Scholar PubMed PubMed Central

[34] T. Nguyen, Y. Xue, Y. Li, L. Tian, and G. Nehmetallah, “Deep learning approach for Fourier ptychography microscopy,” Opt. Express, vol. 26, pp. 26470–26484, 2018. https://doi.org/10.1364/oe.26.026470.Search in Google Scholar PubMed

[35] Y. Wu, Y. Rivenson, H. Wang, et al.., “Three-dimensional virtual refocusing of fluorescence microscopy images using deep learning,” Nat. Methods, vol. 16, pp. 1323–1331, 2019. https://doi.org/10.1038/s41592-019-0622-5.Search in Google Scholar PubMed

[36] Y. Rivenson, Y. Zhang, H. Günaydın, D. Teng, and A. Ozcan, “Phase recovery and holographic image reconstruction using deep learning in neural networks,” Light Sci. Appl., vol. 7, p. 17141, 2018. https://doi.org/10.1038/lsa.2017.141.Search in Google Scholar PubMed PubMed Central

[37] Y. Wu, Y. Rivenson, Y. Zhang, et al.., “Extended depth-of-field in holographic imaging using deep-learning-based autofocusing and phase recovery,” Optica, vol. 5, p. 704, 2018. https://doi.org/10.1364/optica.5.000704.Search in Google Scholar

[38] Y. Rivenson, Y. Wu, and A. Ozcan, “Deep learning in holography and coherent imaging,” Light Sci. Appl., vol. 8, pp. 1–8, 2019. https://doi.org/10.1038/s41377-019-0196-0.Search in Google Scholar PubMed PubMed Central

[39] Y. Jo, H. Cho, S. Y. Lee, et al.., “Quantitative phase imaging and artificial intelligence: a review,” IEEE J. Sel. Top. Quantum Electron., vol. 25, pp. 1–14, 2019. https://doi.org/10.1109/jstqe.2018.2859234.Search in Google Scholar

[40] Y. Park, C. Depeursinge, and G. Popescu, “Quantitative phase imaging in biomedicine,” Nat. Photonics, vol. 12, pp. 578–589, 2018. https://doi.org/10.1038/s41566-018-0253-x.Search in Google Scholar

[41] N. Goswami, Y. R. He, Y.-H. Deng, et al.., “Label-free SARS-CoV-2 detection and classification using phase imaging with computational specificity,” Light Sci. Appl., vol. 10, p. 176, 2021. https://doi.org/10.1038/s41377-021-00620-8.Search in Google Scholar PubMed PubMed Central

[42] B. Rahmani, D. Loterie, G. Konstantinou, D. Psaltis, and C. Moser, “Multimode optical fiber transmission with a deep learning network,” Light Sci. Appl., vol. 7, pp. 1–11, 2018. https://doi.org/10.1038/s41377-018-0074-1.Search in Google Scholar PubMed PubMed Central

[43] N. Borhani, E. Kakkava, C. Moser, and D. Psaltis, “Learning to see through multimode fibers,” Optica, vol. 5, p. 960, 2018. https://doi.org/10.1364/optica.5.000960.Search in Google Scholar

[44] V. Bianco, P. L. Mazzeo, M. Paturzo, C. Distante, and P. Ferraro, “Deep learning assisted portable IR active imaging sensor spots and identifies live humans through fire,” Opt. Lasers. Eng., vol. 124, p. 105818, 2020. https://doi.org/10.1016/j.optlaseng.2019.105818.Search in Google Scholar

[45] S. You, E. J. Chaney, H. Tu, Y. Sun, S. Sinha, and S. A. Boppart, “Label-free deep profiling of the tumor microenvironment,” Cancer Res., vol. 81, pp. 2534–2544, 2021. https://doi.org/10.1158/0008-5472.can-20-3124.Search in Google Scholar

[46] J. Yoon, Y. Jo, M.-H. Kim, et al.., “Identification of non-activated lymphocytes using three-dimensional refractive index tomography and machine learning,” Sci. Rep., vol. 7, p. 6654, 2017. https://doi.org/10.1038/s41598-017-06311-y.Search in Google Scholar PubMed PubMed Central

[47] J. Li, J. Garfinkel, X. Zhang, et al.., “Biopsy-free in vivo virtual histology of skin using deep learning,” Light Sci. Appl., vol. 10, p. 233, 2021. https://doi.org/10.1038/s41377-021-00674-8.Search in Google Scholar PubMed PubMed Central

[48] Y. Rivenson, T. Liu, Z. Wei, Y. Zhang, K. de Haan, and A. Ozcan, “PhaseStain: the digital staining of label-free quantitative phase microscopy images using deep learning,” Light Sci. Appl., vol. 8, p. 23, 2019. https://doi.org/10.1038/s41377-019-0129-y.Search in Google Scholar PubMed PubMed Central

[49] J. Wu, L. Cao, L. Cao, G. Barbastathis, and G. Barbastathis, “DNN-FZA camera: a deep learning approach toward broadband FZA lensless imaging,” Opt. Lett., vol. 46, pp. 130–133, 2021. https://doi.org/10.1364/ol.411228.Search in Google Scholar

[50] A. Sinha, J. Lee, S. Li, and G. Barbastathis, “Lensless computational imaging through deep learning,” Optica, vol. 4, p. 1117, 2017. https://doi.org/10.1364/optica.4.001117.Search in Google Scholar

[51] Y. Rivenson, H. Wang, Z. Wei, et al.., “Virtual histological staining of unlabelled tissue-autofluorescence images via deep learning,” Nat. Biomed. Eng., vol. 3, pp. 466–477, 2019. https://doi.org/10.1038/s41551-019-0362-y.Search in Google Scholar PubMed

[52] I. Malkiel, M. Mrejen, A. Nagler, U. Arieli, L. Wolf, and H. Suchowski, “Plasmonic nanostructure design and characterization via Deep Learning,” Light Sci. Appl., vol. 7, p. 60, 2018. https://doi.org/10.1038/s41377-018-0060-7.Search in Google Scholar PubMed PubMed Central

[53] W. Ma, F. Cheng, and Y. Liu, “Deep-learning-Enabled on-demand design of chiral metamaterials,” ACS Nano, vol. 12, pp. 6326–6334, 2018. https://doi.org/10.1021/acsnano.8b03569.Search in Google Scholar PubMed

[54] W. Ma, F. Cheng, Y. Xu, Q. Wen, and Y. Liu, “Probabilistic representation and inverse design of metamaterials based on a deep generative model with semi-supervised learning strategy,” Adv. Mater., vol. 31, p. 1901111, 2019. https://doi.org/10.1002/adma.201901111.Search in Google Scholar PubMed

[55] E. Schwartz, R. Giryes, and A. M. Bronstein, “DeepISP: towards learning an end-to-end image processing pipeline,” IEEE Trans. Image Process, vol. 28, pp. 912–923, 2019. https://doi.org/10.1109/tip.2018.2872858.Search in Google Scholar

[56] V. Sitzmann, S. Diamond, Y. Peng, et al.., “End-to-end optimization of optics and image processing for achromatic extended depth of field and super-resolution imaging,” ACM Trans. Graph., vol. 37, pp. 1–13, 2018. https://doi.org/10.1145/3197517.3201333.Search in Google Scholar

[57] Z. Ballard, C. Brown, A. M. Madni, and A. Ozcan, “Machine learning and computation-enabled intelligent sensor design,” Nat. Mach. Intell., vol. 3, pp. 556–565, 2021. https://doi.org/10.1038/s42256-021-00360-9.Search in Google Scholar

[58] H.-A. Joung, Z. S. Ballard, J. Wu, et al.., “Point-of-Care serodiagnostic test for early-stage lyme disease using a multiplexed paper-based immunoassay and machine learning,” ACS Nano, vol. 14, pp. 229–240, 2020. https://doi.org/10.1021/acsnano.9b08151.Search in Google Scholar PubMed

[59] C. A. Metzler, H. Ikoma, Y. Peng, and G. Wetzstein, “Deep optics for single-shot high-dynamic-range imaging,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, 2020, pp. 1372–1382.10.1109/CVPR42600.2020.00145Search in Google Scholar

[60] L. Hu, S. Hu, W. Gong, and K. Si, “Learning-based Shack-Hartmann wavefront sensor for high-order aberration detection,” Opt. Express, vol. 27, pp. 33504–33517, 2019. https://doi.org/10.1364/oe.27.033504.Search in Google Scholar

[61] M. Schmeisser, B. C. Heisen, M. Luettich, et al.., “Parallel, distributed and GPU computing technologies in single-particle electron microscopy,” Acta Crystallogr. D Biol. Crystallogr., vol. 65, pp. 659–671, 2009. https://doi.org/10.1107/s0907444909011433.Search in Google Scholar PubMed PubMed Central

[62] B. J. Shastri, A. N. Tait, T. F. de Lima, M. A. Nahmias, H.-T. Peng, and P. R. Prucnal, “Neuromorphic photonics, principles of,” in Encyclopedia of Complexity and Systems Science, R. A. Meyers, Ed., Berlin Heidelberg, Springer, 2018, pp. 1–37.10.1007/978-3-642-27737-5_702-1Search in Google Scholar

[63] Y. Shen, N. C. Harris, S. Skirlo, et al.., “Deep learning with coherent nanophotonic circuits,” Nat. Photonics, vol. 11, pp. 441–446, 2017. https://doi.org/10.1038/nphoton.2017.93.Search in Google Scholar

[64] A. N. Tait, T. F. de Lima, E. Zhou, et al.., “Neuromorphic photonic networks using silicon photonic weight banks,” Sci. Rep., vol. 7, no. 1, p. 7430, 2017. https://doi.org/10.1038/s41598-017-07754-z.Search in Google Scholar PubMed PubMed Central

[65] X. Xu, M. Tan, B. Corcoran, et al.., “11 TOPS photonic convolutional accelerator for optical neural networks,” Nature, vol. 589, pp. 44–51, 2021. https://doi.org/10.1038/s41586-020-03063-0.Search in Google Scholar PubMed

[66] J. Feldmann, N. Youngblood, C. D. Wright, H. Bhaskaran, and W. H. P. Pernice, “All-optical spiking neurosynaptic networks with self-learning capabilities,” Nature, vol. 569, pp. 208–214, 2019. https://doi.org/10.1038/s41586-019-1157-8.Search in Google Scholar PubMed PubMed Central

[67] A. N. Tait, M. A. Nahmias, B. J. Shastri, and P. R. Prucnal, “Broadcast and weight: an integrated network for scalable photonic spike processing,” J. Lightwave Technol., vol. 32, pp. 4029–4041, 2014. https://doi.org/10.1109/jlt.2014.2345652.Search in Google Scholar

[68] A. N. Tait, T. F. de Lima, M. A. Nahmias, et al.., “Silicon photonic modulator neuron,” Phys. Rev. Appl., vol. 11, p. 064043, 2019. https://doi.org/10.1103/physrevapplied.11.064043.Search in Google Scholar

[69] G. Wetzstein, A. Ozcan, S. Gigan, et al.., “Inference in artificial intelligence with deep optics and photonics,” Nature, vol. 588, pp. 39–47, 2020. https://doi.org/10.1038/s41586-020-2973-6.Search in Google Scholar PubMed

[70] X. Lin, Y. Rivenson, N. T. Yardimci, et al.., “All-optical machine learning using diffractive deep neural networks,” Science, vol. 361, pp. 1004–1008, 2018. https://doi.org/10.1126/science.aat8084.Search in Google Scholar PubMed

[71] O. Kulce, D. Mengu, Y. Rivenson, and A. Ozcan, “All-optical information-processing capacity of diffractive surfaces,” Light Sci. Appl., vol. 10, p. 25, 2021. https://doi.org/10.1038/s41377-020-00439-9.Search in Google Scholar PubMed PubMed Central

[72] E. Goi, X. Chen, Q. Zhang, et al.., “Nanoprinted high-neuron-density optical linear perceptrons performing near-infrared inference on a CMOS chip,” Light Sci. Appl., vol. 10, p. 40, 2021. https://doi.org/10.1038/s41377-021-00483-z.Search in Google Scholar PubMed PubMed Central

[73] O. Kulce, D. Mengu, Y. Rivenson, and A. Ozcan, “All-optical synthesis of an arbitrary linear transformation using diffractive surfaces,” Light Sci. Appl., vol. 10, p. 196, 2021. https://doi.org/10.1038/s41377-021-00623-5.Search in Google Scholar PubMed PubMed Central

[74] C. Isil, D. Mengu, Y. Zhao, et al.., “Super-resolution image display using diffractive decoders,” arXiv, 2022. https://doi.org/10.48550/ARXIV.2206.07281.Search in Google Scholar

[75] D. Mengu and A. Ozcan, “All-optical phase recovery: diffractive computing for quantitative phase imaging,” Adv. Opt. Mater., p. 2200281, 2022. https://doi.org/10.1002/adom.202200281.Search in Google Scholar

[76] Y. Luo, Y. Zhao, J. Li, et al.., “Computational imaging without a computer: seeing through random diffusers at the speed of light,” eLight, vol. 2, p. 4, 2022. https://doi.org/10.1186/s43593-022-00012-4.Search in Google Scholar

[77] D. Mengu, Y. Zhao, N. T. Yardimci, Y. Rivenson, M. Jarrahi, and A. Ozcan, “Misalignment resilient diffractive optical networks,” Nanophotonics, vol. 9, pp. 4207–4219, 2020. https://doi.org/10.1515/nanoph-2020-0291.Search in Google Scholar

[78] D. P. Kingma and J. Ba, “Adam: a method for stochastic optimization,” in 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, May 7-9, 2015, Conference Track Proceedings, Y. Bengio and Y. LeCun, Eds., 2015.Search in Google Scholar


Supplementary Material

The online version of this article offers supplementary material (https://doi.org/10.1515/nanoph-2022-0358).


Received: 2022-06-21
Accepted: 2022-08-23
Published Online: 2022-09-05

© 2022 the author(s), published by De Gruyter, Berlin/Boston

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 29.11.2023 from https://www.degruyter.com/document/doi/10.1515/nanoph-2022-0358/html
Scroll to top button