Skip to main content
iScience logoLink to iScience
. 2025 Mar 14;28(4):112222. doi: 10.1016/j.isci.2025.112222

Optical multilayer thin film structure inverse design: From optimization to deep learning

Taigao Ma 1,3, Mingqian Ma 2, L Jay Guo 2,
PMCID: PMC11995089  PMID: 40230531

Summary

Optical multilayer thin film structures have been widely used in numerous photonic domains and applications. The key component to enable these applications is the inverse design. Different from other photonic structures such as metasurface or waveguide, multilayer thin film is a one-dimensional structure, which deserves its own treatment for the design process. Optimization has always been the standard design algorithm for decades. Recent years have witnessed a rapid development of integrating different deep learning algorithms to tackle the inverse design problems. A natural question to ask is: how do these algorithms differ from each other? Why do we need to develop so many algorithms and what type of challenges do they solve? What is the state-of-the-art algorithm in this domain? Here, we review recent progress and provide a guide tour through this research area, starting from traditional optimization to recent deep learning approaches. Challenges and future perspectives are also discussed.

Subject area: Natural sciences, Physics, Applied sciences

Graphical abstract

graphic file with name fx1.jpg


Natural sciences; Physics; Applied sciences

Introduction

Optical multilayer thin film structure1,2,3 (shortened as multilayer structure) is a type of photonic platform that consists of multiple layers of material stacked together, with thickness typically in the range of tens of nanometers to several hundred nanometers. Because of their ease of fabrication, they have been widely used in many different photonic applications. For example, most photovoltaics4,5 devices are based on thin film structures to achieve high power conversion efficiency in large-scale manufacturing. Multilayer thin film-based transmissive/reflective filters6,7 are widely used as functional coatings and protective layers for optical lenses and glass pieces. The light interference phenomenon inside multilayer structures also provides an ideal platform to make structural color coatings8,9 as an eco-friendly and advanced decoration with high aesthetic expression. Other important applications include displays,10,11 touch screens,12 radiative cooling devices,13 to name a few.

For a given multilayer structure, obtaining its optical responses is easy and straightforward, through well-developed electromagnetic simulation algorithms, such as Transfer Matrix Methods14 (TMM). However, the inverse problem, to design a multilayer structure with desired optical responses is non-trivial and usually does not have a closed-form. The inverse design requires determining the total number of layers and identifying the best combination of materials at each layer as well as their corresponding thickness, which is a combinatorial optimization problem. At the early stage, many designs are based on researchers’ physics intuition or optical knowledge. However, this method is only suitable for some simple and conventional structures. As design complexity increases, optimization-based methods have been widely adapted to automatically identify optimal structures, including needle optimization,15,16 particle swarm optimization,17 genetic algorithms,18 and so forth. To speed up the design process, in recent several years, researchers have started to use deep learning as an alternative approach for inverse design.19,20,21,22,23 After learning from the training dataset, these deep learning models can establish a direct mapping from optical responses to the multilayer structure and, therefore, can finish the design process within a fraction of a second, with a speed much faster than the optimization approach. Many different deep learning models have already emerged and demonstrated excellent performance in diverse applications, including tandem networks,19,24,25 Variational Auto-Encoders (VAE),26 Generative Adversarial Networks (GAN),27 Mixture Density Networks (MDN),20,21 and so forth. There are also many nice benchmarking works that compare different inverse design methods.28,29,30

With such significant progress in multilayer structure inverse design, a comprehensive review of these inverse design algorithms could benefit the relevant research community by providing a high-level perspective on the development and future directions. Though there are many excellent reviews that summarize deep learning-based inverse design for photonic and nano-photonic,31,32,33,34,35,36,37,38 there is essentially no review that focuses on the multilayer structure inverse design. Different from other photonic structures such as metasurface or waveguide, multilayer thin film is a one-dimensional structure, which requires special treatment and considerations during the design process. For example, recently many researchers started to treat multilayer structures as a sequence,23,39,40 which is a special type of data that has not been considered in other types of photonic structures.

In this review, we will fill this gap by providing a perspective of multilayer structure inverse design and looking through the research roadmap. In the first part, we will provide an overview of multilayer structure inverse design by dissecting the representations of both multilayer structures and optical responses and present the overall challenges. Then we will go through different mainstream inverse design methods and summarize their achievements from a historical perspective: starting from the traditional optimization-based methods over decades ago, then the deep learning using vectorized representation starting around several years ago, the recent combined methods using both optimization and deep learning, and finally the deep learning using sequential representation, which was just proposed in recent two years. We hope this review could provide insights into current research development and promote the exploration of new research frontiers.

Multilayer structure inverse design

In this section, we will provide a general introduction to multilayer inverse design fundamentals. Inverse design usually deals with finding the suitable structures that can satisfy the desired optical response. Therefore, there are two aspects that researchers are interested in: multilayer structures and optical responses. We will first describe core concepts that we need to figure out when designing a multilayer structure, as well as three different methods when representing a multilayer structure. This is the part where readers can tell the difference between multilayer design and other nanophotonic inverse designs. Following this, several common types of optical responses with general interest and wide applications will be discussed. In the last part of this section, we will summarize several challenges during inverse design and provide an illustrative table to show how existing inverse design methods can tackle these challenges in different aspects. Details of these methods will be discussed in the following sections.

Design parameters for multilayer structure

As the name suggests, a multilayer structure consists of multiple layers of materials stacking on top of each other. Usually, each layer can be made from a different material with different thickness, and these different layers contribute together to control light propagation and interaction inside the medium, and collectively determine the multilayer structure’s optical responses. For a multilayer structure, there are three sets of parameters to be determined during the design process: the total number of layers, material combinations at each layer, and their corresponding thickness.

The total number of layers usually depends on the type of optical responses. For example, only 3–4 layers of suitable materials can already make a nice structural color filter.8,41,42 However, in some other cases, such as designing a perfect reflective or transmissive filter, the total number of layers can easily exceed 100. Generally, there is no specific criterion for how many layers should be used for a given application; however, it is always desirable to use structures with fewer layers considering the fabrication complexity and cost. Material combinations are another important design parameter. In optical frequency, different materials usually have different refractive indices, which uniquely determine how light will pass through and interact with the layers. Based on their absorption ability in optical frequency, most materials can be divided into two types: dielectric, or metallic. Dielectric materials such as silicon dioxide (SiO2) or titanium dioxide (TiO2) typically have low optical absorption and allow a significant portion of light to pass through in the visible range, while metallic materials such as silver (Ag) or gold (Au) have high absorption and are usually opaque in optical frequency. Note that we also include semiconductor materials in the dielectric category. Identifying the correct material selections at each layer can effectively determine how light is propagating inside the layers as well as light refraction at each interface, and thus, the optical responses in the layered system. However, this is not a trivial task as materials are usually dispersive, i.e., the refractive index is dependent on the incident light wavelength. The last important design parameter is the thickness at each layer. The major usage of thickness is to introduce a desirable propagation phase while light passes through each layered medium. In some special cases, this can create constructive or destructive interference phenomenon. One example is the Distributed Bragg Reflector (DBR)43,44,45 that consists of alternating materials with high/low refractive index profiles. When each layer’s thickness is around one-quarter of the effective wavelength, strong interference in the reflection side can lead to a perfect reflector that is close to 100%.

As an example, in Figure 1A, we illustrate a simple five-layer structure (left side) and use a table to illustrate these three design parameters (right side). This structure is made on a glass substrate which is not treated as a design parameter. As a notation, we count the number of layers from top to bottom, i.e., the first layer is closer to the air and the fifth layer is closer to the substrate. We will use this notation for the rest of this review. Ideally, we hope that an inverse design algorithm can identify all three design parameters based on the provided target optical responses. However, there are significant difficulties when trying to incorporate them into a computer-based inverse design algorithm. For example, how can we let an algorithm to understand the number of layers, which is supposed to be a varying number? In addition, how to represent different materials in the algorithm? Remember that materials usually have dispersion, which hinders us from using a single refractive index to distinguish them.

Figure 1.

Figure 1

Visualization of multilayer structures as well as three common types of representation

(A) An example of a five-layer thin film structure (left) as well as the three design parameters presented inside a table (right), including the total number of layers, the material combinations, and their thickness at each layer.

(B–D) Shows three widely used representation for a multilayer structure in inverse design algorithms. (B) is the vectorized representation, where material and total number of layers are fixed and only continuous thickness will be designed. (C) is the combined representation, where total number of layers are fixed and we need to design both material and thickness. Usually, material is a discrete variable that is selected from a pre-defined material database. (D) is the sequential representation, where all three parameters will be designed in a layer-by-layer way.

Perhaps the easiest design parameter to deal with is the thickness, which is a continuous variable that aligns well with computational considerations. Currently, there have been three different ways of representing a multilayer structure inside the inverse design algorithm and the first type is the vectorized representation that only relates to the thickness parameter. We illustrate this idea in Figure 1B. Within this representation, one needs to first determine how many layers to use as well as material choices at each layer based on their prior knowledge or experience, and only use the algorithm to find out the thickness. Common algorithms include some optimization-based methods such as Particle Swarm Optimization17 (PSO), and many machine learning based methods such as tandem networks.19

The major limitation of this vectorized representation comes from the fact that its design performance is heavily determined by the pre-selected materials. Considering the large variety of materials in nature, it is possible that these selected materials may not lead to a good performance for any thickness choices,9 which can leave the design algorithm running in vain. An inverse design algorithm that can automatically determine the material at each layer would be preferable. This leads to the second type of combined representation (see Figure 1C). Here, the material is represented by a discrete variable and thickness is a continuous variable. Usually, there are several candidate materials to be selected and they are denoted by different integers. For example, when there are two materials, we can use 0 to represent one material and 1 for the other material. In some special cases when all candidate materials have low dispersion, a single refractive index value can be used to represent the material. Many global optimization methods use this type of representation, such as genetic algorithm18 and memetic algorithm.46

This combined representation has been popular for a long time. Only very recently, with the development of sequential machine learning models such as Recurrent Neural Networks47,48 (RNN) and transformers,49 researchers started to establish the third idea of sequential representation, describing a multilayer structure with a sequence (see Figure 1D). Instead of splitting material and thickness separately, this sequential representation treats both material and thickness on an equal footing, and assembles a multilayer structure in a layer-by-layer manner, just like in actual fabrication. Two recent models, the sequential decision network called OMLPPO,23,40 and the conditional sequence generation network called OptoGPT,39,50 are adapting this representation. Both methods have been demonstrated to automatically determine the optimal number of layers not known beforehand and terminate the design process when a target is reached.

Optical responses

In the previous section we mainly talk about the output of the inverse design, namely, the design parameters. Here, we would like to briefly summarize the input of inverse design: the targeted optical responses for the designed structure. In optical frequency, multilayer structures have been widely used as transmissive/reflective filters, structural color coatings, absorbers, radiative cooling devices, and so forth. These applications can be attributed to the following types of optical responses.

  • (1)

    Transmission or Reflection Spectrum: They are mostly used as design targets for many special filters. For example, a band-pass/band-notch filter requires a transmission window/block region at the desired frequency. Designing DBR is looking for a near 100% reflection spectrum within the expected wavelength region. On the other hand, a near-zero reflection spectrum is needed for designing anti-reflection coatings.

  • (2)

    Absorption Spectrum: Light absorptance is very important for many energy-harvesting and detection applications. In solar-thermal harvesting devices, a high absorption over a broadband wavelength that covers the whole solar spectrum is always desirable. Many detection devices such as spectrometers require a perfect absorption spectrum to completely remove potential crosstalk and improve the measurement accuracy.

  • (3)

    Structural Color: Compared to traditional color, structural color has unique advantages of higher spatial resolution, better stability, more environmental friendliness, and so forth. They can be well described using a color space such as RGB, xyY, or Lab values.

There are also many other important optical responses not listed here, e.g., combined spectrum and angular responses, which present higher level difficulties. Different from multilayer structures that require special representations in the design algorithm, these optical responses (i.e., spectra) are usually continuous variables and can be easily represented as a high-dimension vector when integrated into the design algorithm. Therefore, we will not provide more description on this.

Challenges

An ideal inverse design algorithm should be able to quickly output multilayer structures given any type of response as the input. However, this can lead to significant difficulties and challenges. The first challenge is how to design the multilayer structure globally (Global Design). By global we mean the algorithm can automatically design for all three sets of design parameters, including the total number of layers, the material sequence, and their thickness at each layer. As a comparison, local design only deals with thickness design, where the performance is heavily constrained by the fixed material selections. Global design is non-trivial as the design space is extremely large. For example, considering a five-layer structure where each layer has ten different material choices and the thickness also only has ten different discrete values, the design space can expand to 1010 different combinations of structures. It is impossible to search all potential designs exhaustively, therefore, advanced algorithms are required to navigate through this large design space and identify the optimal structure.

The second challenge is how to finish a design quickly (Efficient Design). Traditional optimization algorithms are based on iterative evaluations and use fitness functions to update structures to reach user-defined goals. Although many numerical methods such as TMM can finish simulating a multilayer structure and obtaining its optical characteristics within 1 s, the iterative evaluation process often requires thousands or even more simulations, resulting in a significantly long processing time. The situation can be worse when multiple design targets are required. Speeding up the design process and making it more efficient would benefit everyone in this field.

In addition to these two major challenges related to the algorithm itself, researchers and engineers are also interested in leveraging inverse design algorithms to assist the practical fabrication process. Based on this, the third challenge is how to obtain multiple different designed structures with minimal effort so that researchers can select the best one for their fabrication needs based on the availability of materials and deposition methods (Diverse Design). In principle, this is feasible because of the one-to-many mapping feature in the inverse design, meaning multiple structures can lead to similar optical properties. However, most existing algorithms cannot tackle this challenge because they can only output one definite design. An algorithm with probabilistic output can potentially solve this difficulty. The fourth challenge is how to integrate practical constraints into the design process (Flexible Design). With this challenge solved, researchers may restrict the material selections and thickness range at any desired layer based on their fabrication or design needs. For example, some may require the top layer to be dielectric material to protect the next metallic layers from oxidation. The latter two challenges have not yet been extensively explored.

In this review, we will focus on these four challenges and summarize how researchers came up with innovative ideas and advanced algorithms to tackle each of them. Decades ago, some global optimizations, including genetic algorithms18 and memetic algorithms,46 have been proposed to simultaneously design materials and thickness from a global perspective. Later around 2017, with the great success of machine learning, many different neural network-based methods were utilized to speed up the inverse design process. However, both methods have their own disadvantages: optimization-based algorithms are not efficient, while deep learning based algorithms cannot design structures globally. Therefore, later around 2020, the combination of optimization and deep learning provides a solution to alleviate these problems to design in the global space while still making the algorithms fast. However, the iterative process still takes quite a long time. Finally, in recent two or three years, with the development of sequential learning, researchers started to treat the multilayer structure as a sequence and eventually developed OptoGPT that can solve all four challenges, declaring the victory of the long-lasting design issues in multilayer structure. In Table 1, we summarize all these impactful algorithms and list their accomplishments.

Table 1.

A comparison of different inverse design algorithm for multilayer structure

Methods Reference Year Multilayer Representation Global Design? Efficient Design? Diverse Design? Flexible Design?
Needle Optimization Tikhonravov et al.16 2007 Combined Yes No Yes No
Genetic Algorithm Schubert et al.18 2008 Combined Yes No No No
PSO Rabady and Ababneh17 2014 Vectorized No No No No
Memetic Algorithm Shi et al.46 2018 Combined Yes No No No
Tandem Networks Liu et al.19 2018 Vectorized No Yes No No
MDN Unni et al.20 2020 Vectorized No Yes Yes No
VAE Zandehshahvar et al.26 2021 Vectorized No Yes No No
GAN Dai et al.51 2022 Vectorized No Yes Yes No
GLOnet Jiang and Fan52 2021 Combined Yes No Yes No
NEUTRON Wang and Guo53 2022 Combined Yes No Yes No
OMLPPO Wang et al.23 2021 Sequential Yes No Yes No
OptoGPT Ma et al.39 2024 Sequential Yes Yes Yes Yes

Optimization-based methods

Since decades ago, heuristic optimization-based methods have become essential tools in tackling inverse design challenges, offering robust and flexible approaches to explore the vast solution space and find near-optimal designs. Although dealing with inverse design, these methods are mainly based on iterative forward simulation and follow a similar pipeline shown in Figure 2. When given a design target, the algorithm will first start from a random structure and obtain its simulated optical responses. Then, the difference between the simulated responses and target responses will be evaluated and used to calculate the fitness function. If the fitness function does not meet the termination criteria, the algorithm will update the initial structure and keep running this iterative process until reaching satisfaction. After sufficient iterations, a structure with desired optical responses can be found. Different optimization methods, including needle optimization, particle swarm optimization, genetic algorithms, differ from each other in the way how they update the structure in each iteration, with illustrations can be found in Figure 2. A summary and benchmark of optimization-based methods on photonic applications can be found in Schneider et al.30 and Campbell et al.38

Figure 2.

Figure 2

A general pipeline of optimization-based methods for inverse design

Optimization-based methods mainly deal with thickness or material design. In the beginning, researchers use the vectorized representation to only optimize the thickness. One such example is Particle swarm optimization (PSO), an evolutionary algorithm inspired by the social behavior of birds. PSO maintains a set of particles that move through the solution space and are influenced by their own best-known position and the best-known positions of the entire particles. In this case, each particle is a vector of layered thickness that can adjust its trajectories based on its own experience and the experience of neighboring particles, balancing the exploration and exploitation. Rabady et al.17 used PSO to design band-pass filters and demonstrated good performance (see Figure 3A). PSO has also been applied in various real-case multilayer thin film applications, including transmissive spectrum design,54 inverse extraction of thickness and optical constants,55 transmissive filters for image differentiation,56 and so forth.

Figure 3.

Figure 3

Several design examples for optimization-based methods

(A) Design performance of Particle swarm optimization (PSO) algorithm for a band-pass optical interference filter in the range of 1000–2000 nm wavelengths. Reprinted from Rabady and Ababneh17 Copyright @ 2014, Elsevier.

(B) A demonstration of versatile design using Needle Optimization. Reprinted from Tikhonravov et al.16 Copyright @ 2007, Optica Publishing Group.

(C) Genetic Algorithm designing a rejection filter of 90% over the 0.55–0.65 μm region. Reprinted from Schubert et al.18 Copyright @ 2008, Optica Publishing Group.

(D) Iterative refinement of memetic algorithm with mixed-integer programming (MIP) to design better generations. Reprinted from Shi et al.46 Copyright @ 2018, ACS Publications.

To further explore the global design of both material and thickness, Tikhonravov et al.15 introduced needle optimization for optical coating design. Needle Optimization iteratively adds or adjusts layers in a thin film design to improve its optical properties toward the design target by modifying the thickness or refractive index of the material configuration (see Figure 3A). The needle optimization evolves gradually, allowing for the generation of multiple solutions with nearly the same performance metrics, providing the designers with a range of options to select the most practical and manufacturable design. Another example is the application of genetic algorithms (GAs) to the design of optical multilayer structures by Martin et al.18 Genetic algorithms are stochastic optimization methods inspired by the principles of natural selection and genetics. They operate by generating an initial population of potential solutions and iteratively evolving these solutions through selection, crossover, and mutation operations to maximize a fitness function, which is usually taken as the merit function commonly used in traditional approaches. In each iteration, the individuals with the highest fitness score are more likely to be selected for reproduction, ensuring that their genetic material is passed on to the next generation. Crossover combines parts of two parent solutions to create offspring with potentially better performance, while mutation introduces random changes to individual solutions to maintain diversity within the population and prevent premature convergence to local optima. The authors demonstrated the effectiveness of this algorithm by designing the refractive index and thickness of each structure layer for three different optical filters: antireflection coating, beam splitter, and rejection filter (see Figure 3C). This idea has also been applied to the design of broadband reflectors,57 edge filters,58 and so forth.

However, using the refractive index to represent materials restricts the design target to be a narrow band because of materials’ dispersion. To design for a broader wavelength, Shi et al.46 introduced a mixed-integer programming (MIP) with a genetic algorithm solution to optimize the design of multilayer thin films. MIP can handle both discrete and continuous variables in the optimization process. In this context, the discrete variables represent different dielectric materials, while continuous variables correspond to the thickness of each material layer. This algorithm improves upon traditional genetic algorithms by introducing discrete materials into the crossover and mutation steps (see Figure 3C). The authors demonstrated state-of-the-art performance using the algorithm on two design tasks with a broad wavelength range: radiative cooling devices and incandescent light bulb filters. The GA with the MIP algorithm’s ability to design a variety of solutions with distinct material and thickness permutations discovered innovative designs for multilayer thin films at a global solution space for wide wavelengths, materials, and thicknesses.

Deep learning methods with vectorized representation

The fundamental unit inside neural networks is the artificial neuron. As shown in Figure 4A, a single neuron first takes one or multiple inputs, then sums up all these inputs multiplied by learnable weights and outputs results through a nonlinear activation function. In this way, information can be passed in, get processed, and propagated to the next level. Usually, a single neural network (NN) can contain thousands or millions of neurons, leading to millions or billions of learnable weights. After training on a large dataset, these learnable weights are updated to capture the general mapping between the inputs and outputs and process extremely complicated information such as image classification,59,60,61 speech recognition,62,63,64 sequence decision,65,66,67 language understanding,49,68,69,70 and so forth. Different types of neural networks differ in the way how the neurons are connected, and therefore, are suitable for processing different types of data. For example, Convolutional Neural Networks (CNN)59 use the convolutional layers to capture spatial correlations, thus widely used for processing and analyzing image-type data. Recurrent Neural Networks (RNN)47 employ recurrent connections to obtain temporal dependencies, making them well-suited for sequential data.

Figure 4.

Figure 4

Illustration of different types of neural networks

(A–F) Visualizations of (A) a single neuron, (B) Multilayer-Perceptron (MLP), (C) tandem networks, (D) Variational Auto-Encoders (VAE), (E) Generative Adversarial Networks (GAN), and (F) Mixture Density Networks (MDN).

In the previous discussion, we concluded that most of these optical responses (inputs) can be denoted as vectors while multilayer structures (outputs) can be denoted using either vectorized, combined, or sequential representations. However, combined representations require a discretized output for material design, which is non-trivial for NN since their outputs are continuous variables. Also, it was not until recent two or three years ago that researchers figured out the sequential representation. Therefore, most of the early works that used deep learning for inverse design seek to solve the vector-to-vector mapping. A general working process is to first generate a training dataset containing multilayer structures and corresponding optical responses using simulation (e.g., TMM), then train the NN model to fit the inverse mapping function from optical responses to multilayer structures. Once trained, the model can be used to design multilayer structures for different optical responses instantaneously, providing a much faster alternative than optimizations, which usually take minutes or hours.

In terms of the deep learning models for vector-to-vector mapping, perhaps the most straightforward way is the Multilayer-Layer Perceptron (MLP), where neurons are stacked layer-by-layer and all neurons are fully connected to other neurons in nearby layers (see Figure 4B). However, directly using MLP to inverse design will give inaccurate results. Because of the one-to-many mapping issue, i.e., there are multiple possible structures for a given optical response, it is difficult to minimize the loss function (the difference between model outputs and target outputs) and have the NN model converged during training. Eventually, the model will learn to output an averaged structure, which usually does not have the desired optical responses.34 Therefore, special buildup of neural networks is required to handle this one-to-many mapping issue. Many different types of neural networks have been proposed and demonstrated effectiveness in inverse design. Based on the type of outputs, they can be classified into two different categories: (1) deterministic NN, mainly including tandem networks, and (2) stochastic NN (or generative NN), including Generative Adversarial Networks (GAN), Variational Auto-Encoders (VAE), and Mixture Density Networks (MDN). In the remaining section, we are going to discuss each model and summarize its variants and applications.

Deterministic NN

In deterministic NN, the model always maps the input into a single output. The most popular one is the tandem network. As the name “tandem” suggests, tandem networks connect two NN in series (see Figure 4C): the Forward Neural Networks (FNN) and the Inverse Neural Networks (INN). The FNN takes in the structural parameters and outputs the predictions of their corresponding optical responses. Once trained, we can use FNN to quickly predict the optical responses for a given structure with high accuracy. On the other hand, INN takes in the target optical responses and outputs predicted structures. The training of tandem networks involves a two-step training procedure for each individual network. In the first step, we need to train FNN separately to obtain a fast surrogate predictor for optical responses. The second step is to connect the output of INN to this pre-trained FNN and use FNN to supervise the learning of INN. In this way, we can resolve the one-to-many mapping issue by only predicting one structure suggested by FNN.

Tandem networks is first introduced in Liu et al.19 to design transmission spectra for multilayer structures. The multilayer system in consideration is a twenty-layer structure with SiO2/Si3N4 alternating material combinations. Using this method, the authors have demonstrated excellent design performance for arbitrary spectra, as well as the special Gaussian shaped spectra (see Figure 5A). In addition, it only takes a fraction of a second to finish a design, which is much faster than optimization-based methods. Later, extensive works have been dedicated to utilizing tandem networks to design different multilayer structures and tackle various design problems efficiently, including designing a Fabry-Perot-cavity-based color filter,71 organic light-emitting diode,72 and so forth. In addition to multilayer structures, tandem networks are also widely used in other types of photonic structures and target responses.73,74,75,76 For example, Gao et al.73 used tandem networks to design reflective structural color for silicon nanorod metasurface; Ma et al.74 combined tandem networks to design a reflection spectrum for three-dimensional chiral metamaterials; He et al.75 utilized tandem networks to design the far-field radiation profile for plasmonic nanoparticles.

Figure 5.

Figure 5

Deterministic NN for multilayer structure inverse design

(A) Tandem network visualization in Liu et al.19 as well as their design performance on gaussian-shaped spectrum. Reprinted from Liu et al.19 Copyright @ 2018, ACS Publications.

(B) Performance comparison of original tandem network with modified tandem network using ReLU activation function with intermediate structural loss. Reprinted from Xu et al.24 Copyright @ 2021, Elsevier.

(C) A modified tandem network: meta-material spectrum transformer (MST), which uses transformer as the building block for both forward and inverse networks. The model achieves good performance on solar metamaterial absorber (SMA). Reprinted from Chen et al.22 Copyright @ 2023, Wiley.

There are also many variants to improve the design performance of tandem networks. For example, Xu et al.24 modified and benchmarked different activation functions and loss functions used in tandem networks to find the optimal schema. They found that using ReLU activation function77 together with intermediate structural loss can significantly increase the design accuracy (see Figure 5B). On the other hand, many researchers seek to find alternate neural networks to replace the MLP architecture used in INN and FNN.22,78 For example, Chen et al.22 proposed to use transformer architecture for both INN and FNN by leveraging the self-attention mechanism to effectively extract and process information. In this work, they proposed a meta-material spectrum transformer (MST) to design a high-performance broadband solar metamaterial absorber that exhibits a high average absorptance of 94% in a broad solar spectrum (see Figure 5C). In addition, considering that the original tandem network can only output one designed structure, Yuan et al.25 came up with the idea of connecting the INN to multiple different FNNs, which they called multi-head tandem networks, to expand its capability of outputting multiple structures.

Stochastic NN

Different from deterministic NN, stochastic NN (also called generative NN) can output multiple different predictions given the same input. This is because these models take in one latent variable randomly sampled from a normal distribution as an extra input, which enables them to generate different outputs conditioned on this extra input. Therefore, they can intrinsically solve the one-to-many mapping issue. There are three major types of generative NN that are widely used in inverse design: VAE,79 GAN,80 and MDN.81 We illustrate their model architectures in Figure 4D−4F, respectively. There are already many reviews32,33,35 and benchmark28,29,82,83 that discuss these models, but toward a more general audience on the photonic inverse designs. Here, we mainly summarize the work done for multilayer structure designs.

Figure 4D shows the model architecture for VAE, which originates from information theory: it first uses an NN (called encoders) to encode the high-dimensional input information into a low-dimension hidden representation, usually with a normal distribution, then uses another NN (decoders) to decode hidden representations back into the corresponding output. During inverse design, each input (target optical responses) will be encoded to different normal distributions with different means and variances. Sampling from this distribution and sending it through the decoder will give the designed structure. Since this sampling process is random, VAE can output multiple structures when decoding from different latent variables, thus solving the one-to-many mapping issue naturally. For example, Zandehshahvar et al.26 used the VAE to design multilayer structures composed of consecutive layers of SiO2 and TiO2. In their work, they demonstrated that when given a single transmission spectrum as input, their model can output three different structures (Figure 6A left side). The simulation shows that their simulated spectra are close to the target spectrum (Figure 6A right side). VAE has also demonstrated successful applications in other types of photonic structures, including Fano-type resonant nanopillars,84 hybrid plasmonic/phase-change material metasurface,85 a periodic array of Au nanoribbons,86 and so forth.

Figure 6.

Figure 6

Stochastic NN for inverse design with multiple outputs

(A) VAE can output multiple designs that satisfy the desired transmission spectrum. Reprinted from Zandehshahvar et al.26 Copyright @ 2021, Optica Publishing Group.

(B) A three-layer structural color designed by GAN showing different structures. Experimental results also verified the design performance. Reprinted from Dai et al.51 Copyright @ 2022, Elsevier.

(C) The output probability from MDN showing a multi-modality distribution which corresponds to different designs. All designs show desired performance in the transmission spectrum. Reprinted from Unni et al.20 Copyright @ 2020, ACS Publications.

Another type of stochastic NN is GAN (Figure 4E). It also contains two neural networks: the generator NN that generates designed structures based on the random latent variables and the input optical responses, and the critic NN that learns to distinguish if the input structure comes from a model or not. Different from VAE, GAN originated from game theory and these two NN learn to generate by combating with each other: the generator tries to generate outputs that are distributed as close to the training dataset as possible to fool the critic, while the critic tries to distinguish the generated outputs from the ground truth in the training dataset. A successful application is demonstrated in Dai’s work,51 where they used GAN to design a transmissive Fabry-Perot-cavity-based color filter using the three-layer Ag-SiO2-Ag structure and identified an average number of 3.58 solutions for each color. Experimental results in Figure 6B also validated the effectiveness of these designs.

MDN is the third type of stochastic NN (see Figure 4F). Different from VAE and GAN which utilize single-modal normal distribution, MDN models the design parameters as a multimodal probability distribution. Here, each distribution in this multi-modality is parameterized by a different normal distribution with unique mean, variance, and weight. The total number of distributions is a hyperparameter that is usually pre-determined and can be tens or hundreds. After training, the designed structure will be obtained after sampling from this predicted multimodal distribution, which inherently solves the one-to-many mapping issue. As an example, Unni et al.20 utilized MDNs with 16 mixtures to design 10 layers of alternating SiO2 and TiO2 structure. Their results in Figure 6C demonstrated that the MDN can give multiple structures that satisfy the desired transmission spectrum.

All these four different neural networks, including tandem networks, GAN, VAE, and MDN have been exploited for multilayer inverse design with vectorized representation. However, each model still has its own limitations. For example, tandem networks are simple to understand and easy to train, but they can only output one designed structure for one input target, which can lead to difficulty when multiple structures are desired. On the other hand, all the other three models are stochastic models and can output diverse structures. Both GAN and VAE samples from a single-modal distribution, but GAN usually has higher diversity than VAE, while VAE is more robust than GAN. With a multi-modality distribution, MDN can learn a more diverse solution than both GAN and VAE. However, MDN is harder to train because of the increased number of modalities and parameters. A delicate adjustment of a number of modalities based on the specific dataset is required to obtain better performance. For an exhaustive discussion, we recommend some benchmark works that compare these models in a fair setting.28,29,83

Optimization-combined deep learning

Although optimization and deep learning based methods are widely used for inverse design, both have their own strengths and limitations. Optimization has much better flexibility to incorporate the material into design consideration. Designing in this global space with both materials and thicknesses usually leads to better performance than deep learning-based methods. However, optimization requires iterative evaluations to minimize the fitness function, which can be time-consuming. In addition, one needs to restart the optimization process from scratch whenever encountering a new design target, making it task-specific. On the other hand, because deep learning based methods usually use neural networks to learn a general mapping from the space of optical responses to the space of multilayer structure, they can finish a design task much faster than optimization and can quickly adapt to a different design target. The probability output for stochastic NN also enables multiple designs to solve one-to-many mapping issues. However, these neural networks usually have vectorized outputs, making it difficult to incorporate the discrete materials into outputs, thus, resulting in limited design performance. A natural question to ask is if it is possible to combine both methods together and obtain greater benefits. It is encouraging to see that many researchers have already started to do so. Based on how these two methods are integrated together, we can classify them into three categories and we will summarize each of them later in discussion.

Neural networks as a fast surrogate simulator

Neural networks are not only fast for inverse design, but also a fast surrogate simulator to predict the optical responses given the structure.19,50,87 Considering that optimization requires an iterative evaluation process, it is straightforward to use NN as the surrogate simulator to speed up the optimization process. For example, Unni et al.21 proposed a mixture-density-based tandem optimization network to design high reflectors. As shown in Figure 7A, they first pre-trained a forward simulator network to instantly evaluate any designs, then ran optimization algorithms to quickly identify the optimal structure. To further speed up the design process, they trained a mixture-density network to give some initial starting structures for optimization. Using this method, they have successfully designed an ultra-broad high reflector using a 20-layer structure. In another work, Ma et al.88 combined deep learning and multi-objective double annealing algorithms to simultaneously maximize solar spectrum absorption and minimize infrared radiation. As shown in Figure 7B, they first trained an MLP for fast evaluation, then used the numerical optimization to find the multilayer structure that leads to the best performance. Compared to traditional optimization, this deep learning-aided optimization can drastically reduce simulation time and make optimization several orders of magnitude faster.

Figure 7.

Figure 7

Multiple ways to combine optimization and deep learning together

(A) A mixture-density-based tandem optimization network. Reprinted from Unni et al.21 Copyright @ 2021, Elsevier.

(B) Deep learning combined with multi-objective double annealing algorithms. Reprinted from Qu et al.87 Copyright @ 2023, Elsevier.

(C) Global optimization networks (GLOnets) and (D) the design performance comparison using GLOnets and traditional optimization. Reprinted from Jiang and Fan.52 Copyright @ 2021, Elsevier.

(E) NEUTRON: Neural particle swarm optimization for material-aware inverse deign and its design performance combined to NN only (F). Reprinted from Wang and Guo.53 Copyright @ 2022, Cell Press.

Neural network as surrogate gradient update

In addition to being fast, another advantage of neural networks is that they are differentiable. After learning from the dataset, NN can build up a continuous and analytical mapping between input and output spaces, even if the relationship in the original data is too complicated to describe with closed-form equations. Therefore, we can use NN to calculate the gradient information and update the input structures directly. Motivated by this idea, Fouchier et al.89 first trained a forward simulator, then fixed the parameter in NN and used backpropagation to calculate the gradient and update input structure parameters. They successfully used this method to design coatings with different light scattering properties. Another typical work is called global optimization networks (GLOnets).52,90,91 As shown in Figure 7C, GLOnets combine generative NN with a TMM simulator to perform population-based optimization. During the design process, the generative NN will output thickness information and refractive index profile using a probability matrix. The combined multilayer structure output will go through TMM solvers and obtain a loss function, which will be used to calculate the backward gradient and update weights in generative NN. After optimization, the generative NN will map the input optical responses to a designed structure. Using this combined method, they have identified an anti-reflective coating with 1.81% averaged reflection, which is lower than traditional optimization methods (See Figure 7D). In a later work, Zhang et al.92 used GLOnets to design the optical transfer function of a multilayer film at wavelengths of 532 nm and 633 nm and successfully achieved incoherent differentiation with a high resolution of 6.2 um, opening the door for high-speed imaging processing, object tracking, and disease diagnosis.

Two-step design process

Different from previous two methods that use NN as a surrogate model, the third type splits the design process into two steps: first use NN to design material combinations, then use optimization to identify the suitable thickness. An example is the algorithm of Neural Particle Swarm Optimization (NEUTRON), which was introduced by Wang et al.53 to inverse design structural colors. The idea was illustrated in Figure 7E. Here, instead of design material and thickness at the same time, NEUTRON first uses a classifier to predict the most suitable materials at each layer for a given layered structure, then uses MDN to give the initial thickness guess in a probability distribution form and combine PSO to refine and optimize the thickness and achieve optimal performance. Using this two-step process, they demonstrated a much more accurate structural color design compared to single MDN, and used it to output structures for a large number of colors in a painting. We give one design example in Figure 7F.

Deep learning methods with sequential representation

Sequence modeling and understanding is an important research field in machine learning as most of the data in our world can be treated as sequences, including natural language,49 human voice,93 time-series,94 video stream,95 and so forth. Different from vectorized data with pre-determined shapes, sequential data usually have variable length and require special architecture in NN, e.g., the recurrent connection inside RNN,47,48 to handle the serialized information. Recently, researchers have observed that the varying numbers of layers in a multilayer structure is analogous to the varying lengths in sequential data. Therefore, they have started to treat multilayer structures as a sequence and use sequential NN to deal with inverse design. In this way, researchers can completely automate the inverse design process for the first time: NN can design the material and thickness simultaneously at each layer, and can also design the total number of layers by smartly terminating the design process whenever NN thinks the current design is good enough. This points to a bright future for multilayer inverse design. Currently, there are two different sequential learning methods used for inverse design: sequential decision process and conditional sequence generation.

Sequential decision process

The first type of sequential learning model is based on the sequential decision process,66 where the model needs to decide what to do next based on the current situation. It has been widely used in robot manipulations,96 game playing,65,67 self-driving cars,97,98 and so forth. In 2021, Wang et al.23 proposed to use a sequential decision process for multilayer design and introduced the algorithm of OMLPPO based on the reinforcement learning algorithms.66,99 In this work, OMLPPO designs and finalizes the multilayer design in a layer-by-layer manner: after figuring out the material and thickness for the first few layers, it will decide what type of material to use for the next layer as well as its corresponding thickness. This process will keep going until the designed structure reaches the desired performance, or the sequence length (namely, the number of layers) equals the maximum allowable number of layers. The proposed OMLPPO algorithm learns to make sequential decisions in a trial-and-error way and uses iterative evaluation to update designs and approach the ideal structure. The complete pipeline is shown in Figure 8A. Although this iterative process looks similar to heuristic optimizations, OMLPPO can learn multilayer structure design domain knowledge during training and leverage it to improve the design performance. For example, the algorithm automatically designed a five-layer ultra-wideband absorber with 97.64% average absorption, which is ∼2% more than the one designed by human experts (Figure 8B). They also used the algorithm to design incandescent light bulb filters and achieved an 8.5% higher visible light enhancement factor than the structure designed by optimization (Figure 8C). Because of its high performance, OMLPPO has also been adapted to solve other non-trivial design tasks. For example, Saha et al.40 identified two multilayer structures that have a similar visible reflection spectrum as decorative Cr, serving as an environmentally friendly replacement for the harmful and toxic chemical finishing process (Figure 8D).

Figure 8.

Figure 8

Sequential decision process for multilayer design

(A) The working pipeline of OMLPPO algorithm.

(B) The perfect absorption spectrum designed by OMLPPO, which is 2% higher than demonstrated human designs.

(C) The incandescent light bulb filters designed by OMLPPO, which exhibits 8% performance improvement than optimization methods.

(A–C) are reprinted from Wang et al.23 Copyright @ 2021, IOP Publishing.

(D) The decorative Chrome-looking reflective coating designed by OMLPPO. Reprinted from Saha et al.40 Copyright @ 2023, ACS Publications.

Conditional sequence generation

Another type of sequential learning model is conditional sequence generation, where the model needs to generate serialized output based on the input. One important example is the language modeling100: to predict the probability of a sequence of words conditioned on previous words and construct coherent and contextually relevant texts. It is the backbone for a wide range of natural language processing (NLP) applications, including machine translation,101 question answering,102 text generation,103 and so forth. The recent introduction of GPT-type models69,103,104,105 which use the decoder part of the Transformer49 model, not only improves the learning efficiency of language modeling but also enhances the model’s ability to generate more coherent texts by scaling up the model size and training dataset, leading to the development of powerful language models, including ChatGPT,106 Llama,107 BART,70 and so forth. The success of the GPT model in language modeling also intrigues researchers to solve other sequential generation problems in scientific and engineering applications, including DNA sequence generation,108 protein prediction,109 molecule modeling,110,111 and so forth.

Inspired by these successful GPT models, in 2024, Ma et al.39 proposed the OptoGPT model and completely tackled these four challenges for multilayer design. Figure 9A and 9B show the analogy between GPT models and the proposed OptoGPT model. For GPT models, given some language texts as input, e.g., a question or task description, they can generate a sequence of words that are related to the inputs. Similarly, given the input of optical responses (in their case, is reflection and transmission spectrum), the OptoGPT model can output the serialized representation of multilayer structures directly, without any iterative process.

Figure 9.

Figure 9

Idea of conditional sequence generation using OptoGPT model

(A and B) Shows the diagram of general GPT models in NLP and the OptoGPT model in multilayer inverse design, respectively.

(C) Different types of input to the OptoGPT model as the design target.

(D) One example of the “structure tokenization” and “structure serialization” for an N-layer structure on the glass substrate. All figures are reprinted from Ma et al.39 Copyright @ 2024, OE Journals.

To make OptoGPT work, we first introduce the “structure serialization” technique to convert the material and thickness information at each layer into a single token by simply concatenating them together. As shown in Figure 9D, a token is a human-readable notation similar to a word or a phrase in language. Then, we proposed the “structure serialization” technique and added these tokens one by one to form the sequence representation of a multilayer structure. In this way, the OptoGPT model can effectively design for both material and thickness information, and automatically determine the optimal number of layers. The input to the OptoGPT model is optical responses. To make OptoGPT well suited for different types of optical responses, we propose to design for both transmission and reflection spectrum and use multiple techniques to extend the input to diverse responses, including absorption spectrum, reflective and transmissive structural color, and so forth. Examples can be found in Figure 9C.

OptoGPT is a foundation model112 that once trained it can be used to solve extensive downstream tasks and applications to design multilayer thin film structures. OptoGPT demonstrated a unified inverse design toward a range of applications, including transmissive/reflective filter, high reflection filter, perfect absorber, arbitrary absorber, reflective/transmissive structural color, and so forth. Figure 10 gives some inverse design examples. Notably, each design task can be completed within 0.1 s. In addition, because the output is based on probability sampling, therefore, OptoGPT can give different designs by running inference multiple times, which solves the diversity challenges. To solve the flexible design challenge, we further proposed “probability resampling” to design structures that can satisfy arbitrary constraints during fabrication or other practical requirements. Using finetuning and “mixed sampling,” OptoGPT can even design structures for different incident angles and polarization states, as well as simultaneous design for multiple angles requirements. In this way, OptoGPT solves all previously mentioned challenges in multilayer inverse design, significantly pushing the research boundaries forward.

Figure 10.

Figure 10

Examples of different types of optical designs that OptoGPT can design with

(A) Design for a band-notch filter at 550 nm.

(B) Design for high reflection in near-infrared.

(C) Design for perfect absorber.

(D) Design for arbitrary absorber.

(E–F) Design for reflective/transmissive type structural color. All figures are reprinted from Ma et al.39 Copyright @ 2024, OE Journals.

Discussion and outlook

In conclusion, this review summarized the approaches of different types of inverse design algorithms specifically for multilayer thin film structures, starting from traditional optimization-based methods, to the recent popular deep learning based algorithms. Optimization-based methods require iterative simulations of multilayer structures as well as evaluations of the fitness function, which can be time-consuming for complicated structures and multiple designs. On the other hand, deep learning based methods learn a general mapping from some established datasets, therefore, they are much faster than optimization. However, considering that collecting training datasets and training modes could take much time, it would be wise to first estimate the time consumption as compared with the task to be accomplished, and then decide which method to use. In addition, optimization-based methods are more flexible when dealing with new materials from the lab, or some custom multilayer structures with different surrounding mediums and substrates. In both cases, one would need to collect a new dataset and retrain deep learning models.

Compared to other types of photonic structures, multilayer structures are special as they can be represented using three different methods. Each of them has its unique advantages and drawbacks, and corresponds to a specific type of inverse design algorithms. It was only until recently that researchers have identified the sequential representation, a unique data type that is specific to layered structure. This new method has shown promising potential to reshape the inverse design.

Despite the great success deep learning-based methods have achieved, there are still remaining questions to be solved. Perhaps the biggest question too deep learning is the “black-box problem,” i.e., the predictions from NN cannot be properly explained due to the high-level complexity and nonlinearity inside NN. In inverse design, this explainability is very important to understand why certain structures are preferred during the design process, e.g., if the designed structure is coming from some existing physical behaviors or potentially includes a new physical mechanism that has not been discovered before. However, uncovering the black-box problem has not been extensively explored yet. New ideas and approaches are needed to tackle this issue and reveal hidden physical principles.

Another issue is that many deep learning based methods require a large dataset. Although after training, NN is extremely fast for inverse design, generating the training dataset itself can take a long time. This leads to the dilemma of whether the speed-up in the inverse design process is worthwhile. In addition to the data-based deep learning, currently another type of neural operator-based learning has been popular in solving complicated partial differential equations, such as PINN113,114 and DeepONet.115,116 By integrating the physics rules and mathematical equations into NN, these models can directly learn the complicated mapping from input to output without any data. More efforts are needed in order to incorporate these methods into inverse design.

Finally, there are still several issues that require special attention for both methods. For example, it is important to consider the physical conditions for the size of the incoming light beams during the design process, especially for the applications where the input light spot size is very small, such as functional filters for laser beams. If these conditions are not properly incorporated into design considerations, the multilayer structure may not perform as expected at some higher incidence angle. Currently, these issues have not been extensively explored for both optimization and deep learning-based methods. New methods and techniques are required to solve these challenges. One possible direction to take is to treat the specialized cases as “out-of-distribution” datasets, where we have recently proposed a method to handle some of these situations, based on the pre-trained transformer and incorporating perturbations and genetic optimization.117

Acknowledgments

We thank the National Science Foundation FET-2309403 for the support of this work.

Author contributions

T. M. and L. G conceived the review framework. All authors contribute to the writing.

Declaration of interests

The authors declare no competing interests.

References

  • 1.Babeva T., editor. Optical Thin Films and Structures. MDPI - Multidisciplinary Digital Publishing Institute; 2021. [DOI] [Google Scholar]
  • 2.Piegari A., Flory F., editors. Optical Thin Films and Coatings Woodhead Publishing Series in Electronic and Optical Materials. Woodhead Publishing; 2013. Index; pp. 811–835. [DOI] [Google Scholar]
  • 3.Macleod H.A. Fifth edition. CRC Press, Taylor & Francis Group; 2018. Thin-film Optical Filters. [Google Scholar]
  • 4.Morales-Masis M., De Wolf S., Woods-Robinson R., Ager J.W., Ballif C. Transparent Electrodes for Efficient Optoelectronics. Adv. Electron. Mater. 2017;3 doi: 10.1002/aelm.201600529. [DOI] [Google Scholar]
  • 5.He Z., Zhong C., Su S., Xu M., Wu H., Cao Y. Enhanced power-conversion efficiency in polymer solar cells using an inverted device structure. Nat. Photonics. 2012;6:591–595. doi: 10.1038/nphoton.2012.190. [DOI] [Google Scholar]
  • 6.Yang C., Ji C., Shen W., Lee K.-T., Zhang Y., Liu X., Guo L.J. Compact multilayer film structures for ultrabroadband, omnidirectional, and efficient absorption. Acs Photonics. 2016;3:590–596. [Google Scholar]
  • 7.Fink Y., Winn J.N., Fan S., Chen C., Michel J., Joannopoulos J.D., Thomas E.L. A Dielectric Omnidirectional Reflector. Science. 1998;282:1679–1682. doi: 10.1126/science.282.5394.1679. [DOI] [PubMed] [Google Scholar]
  • 8.Rorem B.A., Cho T.H., Farjam N., Lenef J.D., Barton K., Dasgupta N.P., Guo L.J. Integrating Structural Colors with Additive Manufacturing Using Atomic Layer Deposition. ACS Appl. Mater. Interfaces. 2022;14:31099–31108. doi: 10.1021/acsami.2c05940. [DOI] [PubMed] [Google Scholar]
  • 9.Yang Z., Ji C., Liu D., Guo L.J. Enhancing the Purity of Reflective Structural Colors with Ultrathin Bilayer Media as Effective Ideal Absorbers. Adv. Opt. Mater. 2019;7 doi: 10.1002/adom.201900739. [DOI] [Google Scholar]
  • 10.Zhang Q., Li B., Huang S., Nomura H., Tanaka H., Adachi C. Efficient blue organic light-emitting diodes employing thermally activated delayed fluorescence. Nat. Photonics. 2014;8:326–332. doi: 10.1038/nphoton.2014.12. [DOI] [Google Scholar]
  • 11.Jeong C., Park Y.-B., Guo L.J. Tackling light trapping in organic light-emitting diodes by complete elimination of waveguide modes. Sci. Adv. 2021;7 doi: 10.1126/sciadv.abg0355. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Park Y.-B., Lee S., Tobah M., Ma T., Guo L.J. Optimized optical/electrical/mechanical properties of ultrathin metal films for flexible transparent conductor applications: review [Invited] Opt. Mater. Express. 2023;13:304. doi: 10.1364/OME.473277. [DOI] [Google Scholar]
  • 13.Raman A.P., Anoma M.A., Zhu L., Rephaeli E., Fan S. Passive radiative cooling below ambient air temperature under direct sunlight. Nature. 2014;515:540–544. doi: 10.1038/nature13883. [DOI] [PubMed] [Google Scholar]
  • 14.Byrnes S.J. Multilayer optical calculations. arXiv. 2016 doi: 10.48550/arXiv.1603.02720. Preprint at. [DOI] [Google Scholar]
  • 15.Tikhonravov A.V., Trubetskov M.K., DeBell G.W. Application of the needle optimization technique to the design of optical coatings. Appl. Opt. 1996;35:5493–5508. doi: 10.1364/AO.35.005493. [DOI] [PubMed] [Google Scholar]
  • 16.Tikhonravov A.V., Trubetskov M.K., DeBell G.W. Optical coating design approaches based on the needle optimization technique. Appl. Opt. 2007;46:704–710. doi: 10.1364/AO.46.000704. [DOI] [PubMed] [Google Scholar]
  • 17.Rabady R.I., Ababneh A. Global optimal design of optical multilayer thin-film filters using particle swarm optimization. Optik. 2014;125:548–553. [Google Scholar]
  • 18.Schubert M.F., Mont F.W., Chhajed S., Poxson D.J., Kim J.K., Schubert E.F. Design of multilayer antireflection coatings made from co-sputtered and low-refractive-index materials by genetic algorithm. Opt. Express. 2008;16:5290–5298. doi: 10.1364/oe.16.005290. [DOI] [PubMed] [Google Scholar]
  • 19.Liu D., Tan Y., Khoram E., Yu Z. Training deep neural networks for the inverse design of nanophotonic structures. Acs Photonics. 2018;5:1365–1369. [Google Scholar]
  • 20.Unni R., Yao K., Zheng Y. Deep convolutional mixture density network for inverse design of layered photonic structures. ACS Photonics. 2020;7:2703–2712. doi: 10.1021/acsphotonics.0c00630. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Unni R., Yao K., Han X., Zhou M., Zheng Y. A mixture-density-based tandem optimization network for on-demand inverse design of thin-film high reflectors. Nanophotonics. 2021;10:4057–4065. doi: 10.1515/nanoph-2021-0392. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Chen W., Gao Y., Li Y., Yan Y., Ou J.-Y., Ma W., Zhu J. Broadband Solar Metamaterial Absorbers Empowered by Transformer-Based Deep Learning. Adv. Sci. 2023;10 doi: 10.1002/advs.202206718. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Wang H., Zheng Z., Ji C., Jay Guo L. Automated multi-layer optical design via deep reinforcement learning. Mach. Learn Sci. Technol. 2021;2 [Google Scholar]
  • 24.Xu X., Sun C., Li Y., Zhao J., Han J., Huang W. An improved tandem neural network for the inverse design of nanophotonics devices. Opt. Commun. 2021;481 doi: 10.1016/j.optcom.2020.126513. [DOI] [Google Scholar]
  • 25.Yuan X., Wang S., Gu L., Xie S., Ma Q., Guo J. Multi-headed tandem neural network approach for non-uniqueness in inverse design of layered photonic structures. Opt Laser. Technol. 2024;176 doi: 10.1016/j.optlastec.2024.110997. [DOI] [Google Scholar]
  • 26.Zandehshahvar M., Kiarashi Y., Chen M., Barton R., Adibi A. Inverse design of photonic nanostructures using dimensionality reduction: reducing the computational complexity. Opt. Lett. 2021;46:2634–2637. doi: 10.1364/OL.425627. [DOI] [PubMed] [Google Scholar]
  • 27.Baucour A., Kim M., Shin J. Data-driven concurrent nanostructure optimization based on conditional generative adversarial networks. Nanophotonics. 2022;11:2865–2873. doi: 10.1515/nanoph-2022-0005. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28.Ma T., Tobah M., Wang H., Guo L.J. Benchmarking deep learning-based models on nanophotonic inverse design problems. OES. 2022;1:210012. doi: 10.29026/oes.2022.210012. [DOI] [Google Scholar]
  • 29.Ren S., Mahendra A., Khatib O., Deng Y., Padilla W.J., Malof J.M. Inverse deep learning methods and benchmarks for artificial electromagnetic material design. Nanoscale. 2022;14:3958–3969. doi: 10.1039/D1NR08346E. [DOI] [PubMed] [Google Scholar]
  • 30.Schneider P.-I., Garcia Santiago X., Soltwisch V., Hammerschmidt M., Burger S., Rockstuhl C. Benchmarking Five Global Optimization Approaches for Nano-optical Shape Optimization and Parameter Reconstruction. ACS Photonics. 2019;6:2726–2733. doi: 10.1021/acsphotonics.9b00706. [DOI] [Google Scholar]
  • 31.Liu Z., Zhu D., Raju L., Cai W. Tackling Photonic Inverse Design with Machine Learning. Adv. Sci. 2021;8 doi: 10.1002/advs.202002923. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Ma W., Liu Z., Kudyshev Z.A., Boltasseva A., Cai W., Liu Y. Deep learning for the design of photonic structures. Nat. Photonics. 2021;15:77–90. [Google Scholar]
  • 33.Wiecha P.R., Arbouet A., Girard C., Muskens O.L. Deep learning in nano-photonics: inverse design and beyond. Photonics Res. 2021;9:B182–B200. [Google Scholar]
  • 34.Jiang J., Chen M., Fan J.A. Deep neural networks for the evaluation and design of photonic devices. Nat. Rev. Mater. 2020;6:679–700. [Google Scholar]
  • 35.Hegde R.S. Deep learning: a new tool for photonic nanostructure design. Nanoscale Adv. 2020;2:1007–1023. doi: 10.1039/C9NA00656G. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Li Z., Pestourie R., Lin Z., Johnson S.G., Capasso F. Empowering Metasurfaces with Inverse Design: Principles and Applications. ACS Photonics. 2022;9:2178–2192. doi: 10.1021/acsphotonics.1c01850. [DOI] [Google Scholar]
  • 37.Park J., Kim S., Nam D.W., Chung H., Park C.Y., Jang M.S. Free-form optimization of nanophotonic devices: from classical methods to deep learning. Nanophotonics. 2022;11:1809–1845. doi: 10.1515/nanoph-2021-0713. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Campbell S.D., Sell D., Jenkins R.P., Whiting E.B., Fan J.A., Werner D.H. Review of numerical optimization techniques for meta-device design [Invited] Opt. Mater. Express. 2019;9:1842. doi: 10.1364/OME.9.001842. [DOI] [Google Scholar]
  • 39.Taigao M., Wang H., Guo L.J. OptoGPT: A foundation model for inverse design in optical multilayer thin film structures. Opto Electron. Adv. 2024;7 https://www.oejournal.org/article/doi/10.29026/oea.2024.240062 [Google Scholar]
  • 40.Saha A., Ma T., Wang H., Guo L.J. Environmentally Sustainable and Multifunctional Chrome-like Coatings Having No Chromium Designed with Reinforcement Learning. ACS Appl. Mater. Interfaces. 2023;15:28772–28780. doi: 10.1021/acsami.3c02993. [DOI] [PubMed] [Google Scholar]
  • 41.Wang D., Liu Z., Wang H., Li M., Guo L.J., Zhang C. Structural color generation: from layered thin films to optical metasurfaces. Nanophotonics. 2023;12:1019–1081. doi: 10.1515/nanoph-2022-0063. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Feng W.-J., Ji C., Guo L.J. Primary and Secondary Reflective Color Realized by Full-Solution-Processed Multi-Layer Structures. Adv. Opt. Mater. 2023;11 doi: 10.1002/adom.202300456. [DOI] [Google Scholar]
  • 43.Wang S. Principles of distributed feedback and distributed Bragg-reflector lasers. IEEE J. Quantum Electron. 1974;10:413–427. doi: 10.1109/JQE.1974.1068152. [DOI] [Google Scholar]
  • 44.Zhang C., ElAfandy R., Han J. Distributed Bragg Reflectors for GaN-Based Vertical-Cavity Surface-Emitting Lasers. Appl. Sci. 2019;9:1593. doi: 10.3390/app9081593. [DOI] [Google Scholar]
  • 45.Kimura M., Okahara K., Miyamoto T. Tunable multilayer-film distributed-Bragg-reflector filter. J. Appl. Phys. 1979;50:1222–1225. doi: 10.1063/1.326154. [DOI] [Google Scholar]
  • 46.Shi Y., Li W., Raman A., Fan S. Optimization of Multilayer Optical Films with a Memetic Algorithm and Mixed Integer Programming. ACS Photonics. 2018;5:684–691. doi: 10.1021/acsphotonics.7b01136. [DOI] [Google Scholar]
  • 47.Schmidt R.M. Recurrent Neural Networks (RNNs): A gentle Introduction and Overview. arXiv. 2019 doi: 10.48550/arXiv.1912.05911. Preprint at. [DOI] [Google Scholar]
  • 48.Staudemeyer R.C., Morris E.R. Understanding LSTM -- a tutorial into Long Short-Term Memory Recurrent Neural Networks. arXiv. 2019 doi: 10.48550/arXiv.1909.09586. Preprint at. [DOI] [Google Scholar]
  • 49.Vaswani A., Shazeer N., Parmar N., Uszkoreit J., Jones L., Gomez A.N., Kaiser Ł., Polosukhin I. Advances in neural information processing systems. 2017. Attention is all you need; p. 30. [Google Scholar]
  • 50.Ma T., Wang H., Guo L.J. ICML SynS & ML ICML Workshop; 2023. OL-transformer: A Fast and Universal Surrogate Simulator for Optical Multilayer Thin Film Structures. [Google Scholar]
  • 51.Dai P., Sun K., Yan X., Muskens O.L., de Groot C.H.K., Zhu X., Hu Y., Duan H., Huang R. Inverse design of structural color: finding multiple solutions via conditional generative adversarial networks. Nanophotonics. 2022;11:3057–3069. doi: 10.1515/nanoph-2022-0095. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 52.Jiang J., Fan J.A. Multiobjective and categorical global optimization of photonic structures based on ResNet generative neural networks. Nanophotonics. 2020;10:361–369. [Google Scholar]
  • 53.Wang H., Guo L.J. NEUTRON: Neural Particle Swarm Optimization for Material-Aware Inverse Design of Structural Color. iScience. 2022;25 doi: 10.1016/j.isci.2022.104339. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Siabi-Garjan A., Hassanzadeh R. A computational approach for engineering optical properties of multilayer thin films: Particle swarm optimization applied to Bruggeman homogenization formalism. Eur. Phys. J. Plus. 2018;133:419. doi: 10.1140/epjp/i2018-12203-y. [DOI] [Google Scholar]
  • 55.Ruan Z.-H., Yuan Y., Zhang X.-X., Shuai Y., Tan H.-P. Determination of optical properties and thickness of optical thin film using stochastic particle swarm optimization. Sol. Energy. 2016;127:147–158. doi: 10.1016/j.solener.2016.01.027. [DOI] [Google Scholar]
  • 56.Jin C., Yang Y. Transmissive nonlocal multilayer thin film optical filter for image differentiation. Nanophotonics. 2021;10:3519–3525. doi: 10.1515/nanoph-2021-0313. [DOI] [Google Scholar]
  • 57.Abdullah T.A., Md Imran H., Chowdhury S.I., Haque M.A. 2024 IEEE 14th Annual Computing and Communication Workshop and Conference (CCWC) 2024. A Genetic Algorithm Approach for A Broadband Reflector Design; pp. 597–602. [DOI] [Google Scholar]
  • 58.Li D.G., Watson A.C. Proceedings Third International Conference on Computational Intelligence and Multimedia Applications. ICCIMA’99; 1999. Genetic algorithms in optical thin film optimisation design; pp. 86–90. [DOI] [Google Scholar]
  • 59.Krizhevsky A., Sutskever I., Hinton G.E. Advances in Neural Information Processing Systems. Curran Associates, Inc; 2012. ImageNet Classification with Deep Convolutional Neural Networks.https://proceedings.neurips.cc/paper_files/paper/2012/file/c399862d3b9d6b76c8436e924a68c45b-Paper.pdf [Google Scholar]
  • 60.Deng J., Dong W., Socher R., Li L.-J., Li K., Fei-Fei L. 2009 IEEE Conference on Computer Vision and Pattern Recognition. 2009. ImageNet: A large-scale hierarchical image database; pp. 248–255. [DOI] [Google Scholar]
  • 61.He K., Zhang X., Ren S., Sun J. 2015 IEEE International Conference on Computer Vision (ICCV) (Santiago, Chile) 2015. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification; pp. 1026–1034. [DOI] [Google Scholar]
  • 62.Li J. Recent Advances in End-to-End Automatic Speech Recognition. arXiv. 2022 doi: 10.48550/arXiv.2111.01690. Preprint at. [DOI] [Google Scholar]
  • 63.Gaikwad K., Gawali W., Yannawar P. Article:A Review on Speech Recognition Technique. Int. J. Comp. Appl. 2010;10:16–24. [Google Scholar]
  • 64.Gulati A., Qin J., Chiu C.-C., Parmar N., Zhang Y., Yu J., Han W., Wang S., Zhang Z., Wu Y., et al. Conformer: Convolution-augmented Transformer for Speech Recognition. arXiv. 2020 doi: 10.48550/arXiv.2005.08100. Preprint at. [DOI] [Google Scholar]
  • 65.Mnih V., Kavukcuoglu K., Silver D., Rusu A.A., Veness J., Bellemare M.G., Graves A., Riedmiller M., Fidjeland A.K., Ostrovski G., et al. Human-level control through deep reinforcement learning. Nature. 2015;518:529–533. doi: 10.1038/nature14236. [DOI] [PubMed] [Google Scholar]
  • 66.Sutton R.S., Barto A.G. MIT Press; 2018. Reinforcement Learning: An Introduction. [Google Scholar]
  • 67.Mastering the game of Go with deep neural networks and tree search. Nature. 2016;529:484–489. doi: 10.1038/nature16961. https://www.nature.com/articles/nature16961 [DOI] [PubMed] [Google Scholar]
  • 68.Devlin J., Chang M.-W., Lee K., Toutanova K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv. 2018 doi: 10.18653/V1/N19-1423. Preprint at. [DOI] [Google Scholar]
  • 69.Brown T.B., Mann B., Ryder N., Subbiah M., Kaplan J., Dhariwal P., Neelakantan A., Shyam P., Sastry G., Askell A., et al. Language Models are Few-Shot Learners. arXiv. 2020 doi: 10.48550/arXiv.2005.14165. Preprint at. [DOI] [Google Scholar]
  • 70.Lewis M., Liu Y., Goyal N., Ghazvininejad M., Mohamed A., Levy O., Stoyanov V., Zettlemoyer L. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics; 2020. BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension; pp. 7871–7880. [DOI] [Google Scholar]
  • 71.Dai P., Wang Y., Hu Y., de Groot C.H., Muskens O., Duan H., Huang R. Accurate inverse design of Fabry–Perot-cavity-based color filters far beyond sRGB via a bidirectional artificial neural network. Photon. Res. 2021;9:B236. doi: 10.1364/PRJ.415141. [DOI] [Google Scholar]
  • 72.Kim S., Shin J.M., Lee J., Park C., Lee S., Park J., Seo D., Park S., Park C.Y., Jang M.S. Inverse design of organic light-emitting diode structure based on deep neural networks. Nanophotonics. 2021;10:4533–4541. doi: 10.1515/nanoph-2021-0434. [DOI] [Google Scholar]
  • 73.Gao L., Li X., Liu D., Wang L., Yu Z. A bidirectional deep neural network for accurate silicon color design. Adv. Mater. 2019;31 doi: 10.1002/adma.201905467. [DOI] [PubMed] [Google Scholar]
  • 74.Ma W., Cheng F., Liu Y. Deep-learning-enabled on-demand design of chiral metamaterials. ACS Nano. 2018;12:6326–6334. doi: 10.1021/acsnano.8b03569. [DOI] [PubMed] [Google Scholar]
  • 75.He J., He C., Zheng C., Wang Q., Ye J. Plasmonic nanoparticle simulations and inverse design using machine learning. Nanoscale. 2019;11:17444–17459. doi: 10.1039/C9NR03450A. [DOI] [PubMed] [Google Scholar]
  • 76.Xiong J., Shen J., Gao Y., Chen Y., Ou J.-Y., Liu Q.H., Zhu J. Real-Time On-Demand Design of Circuit-Analog Plasmonic Stack Metamaterials by Divide-and-Conquer Deep Learning. Laser Photon. Rev. 2023;17 doi: 10.1002/lpor.202100738. [DOI] [Google Scholar]
  • 77.Eckle K., Schmidt-Hieber J. A comparison of deep networks with ReLU activation function and linear spline-type methods. Neural Netw. 2019;110:232–242. doi: 10.1016/j.neunet.2018.11.005. [DOI] [PubMed] [Google Scholar]
  • 78.Gao Y., Chen W., Li F., Zhuang M., Yan Y., Wang J., Wang X., Dong Z., Ma W., Zhu J. Meta-Attention Deep Learning for Smart Development of Metasurface Sensors. Adv. Sci. 2024;11 doi: 10.1002/advs.202405750. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 79.Kingma D.P., Welling M. Auto-Encoding Variational Bayes. arXiv. 2022 doi: 10.48550/arXiv.1312.6114. Preprint at. [DOI] [Google Scholar]
  • 80.Goodfellow I.J., Pouget-Abadie J., Mirza M., Xu B., Warde-Farley D., Ozair S., Courville A., Bengio Y. Generative Adversarial Networks. arXiv. 2014 doi: 10.48550/arXiv.1406.2661. Preprint at. [DOI] [Google Scholar]
  • 81.Bishop C.M. Mixture density networks. 1994. https://publications.aston.ac.uk/id/eprint/373/
  • 82.Deng, Y., Dong, J., Ren, S., Khatib, O., Soltani, M., Tarokh, V., Padilla, W.J., and Malof, J.M. Benchmarking Data-Driven Surrogate Simulators for Artificial Electromagnetic Materials. 12.
  • 83.Ren S., Padilla W., Malof J. Benchmarking deep inverse models over time, and the neural-adjoint method. arXiv. 2021 doi: 10.48550/arXiv.2009.12919. Preprint at. [DOI] [Google Scholar]
  • 84.Hemmatyar O., Abdollahramezani S., Kiarashinejad Y., Zandehshahvar M., Adibi A. Full color generation with Fano-type resonant HfO 2 nanopillars designed by a deep-learning approach. Nanoscale. 2019;11:21266–21274. doi: 10.1039/C9NR07408B. [DOI] [PubMed] [Google Scholar]
  • 85.Kiarashinejad Y., Abdollahramezani S., Zandehshahvar M., Hemmatyar O., Adibi A. Deep Learning Reveals Underlying Physics of Light–Matter Interactions in Nanophotonic Devices. Adv. Theory Simul. 2019;2 doi: 10.1002/adts.201900088. [DOI] [Google Scholar]
  • 86.Kiarashinejad Y., Abdollahramezani S., Adibi A. Deep learning approach based on dimensionality reduction for designing electromagnetic nanostructures. NPJ Comput. Mater. 2020;6:12. doi: 10.1038/s41524-020-0276-y. [DOI] [Google Scholar]
  • 87.Qu Y., Jing L., Shen Y., Qiu M., Soljačić M. Migrating Knowledge between Physical Scenarios Based on Artificial Neural Networks. ACS Photonics. 2019;6:1168–1174. doi: 10.1021/acsphotonics.8b01526. [DOI] [Google Scholar]
  • 88.Ma W., Chen W., Li D., Liu Y., Yin J., Tu C., Xia Y., Shen G., Zhou P., Deng L., Zhang L. Deep learning empowering design for selective solar absorber. Nanophotonics. 2023;12:3589–3601. doi: 10.1515/nanoph-2023-0291. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Design of multilayer optical thin-films based on light scattering properties and using deep neural networks https://opg.optica.org/oe/fulltext.cfm?uri=oe-29-20-32627&id=459831. [DOI] [PubMed]
  • 90.Jiang J., Fan J.A. Global optimization of dielectric metasurfaces using a physics-driven neural network. Nano Lett. 2019;19:5366–5372. doi: 10.1021/acs.nanolett.9b01857. [DOI] [PubMed] [Google Scholar]
  • 91.Jiang J., Fan J.A. Simulator-based training of generative models for the inverse design of metasurfaces. Nanophotonics. 2019;9:1059–1069. doi: 10.1515/nanoph-2019-0330. [DOI] [Google Scholar]
  • 92.Zhang X., Bai B., Sun H., Jin G., Valentine J. Incoherent Optoelectronic Differentiation Based on Optimized Multilayer Films. Laser Photon. Rev. 2022;16 doi: 10.1002/lpor.202200038. [DOI] [Google Scholar]
  • 93.Latif S., Zaidi A., Cuayahuitl H., Shamshad F., Shoukat M., Qadir J. Transformers in Speech Processing: A Survey. arXiv. 2023 doi: 10.48550/arXiv.2303.11607. Preprint at. [DOI] [Google Scholar]
  • 94.Wen Q., Zhou T., Zhang C., Chen W., Ma Z., Yan J., Sun L. Transformers in Time Series: A Survey. arXiv. 2023 doi: 10.48550/arXiv.2202.07125. Preprint at. [DOI] [Google Scholar]
  • 95.Yaqoob A., Bi T., Muntean G.-M. A Survey on Adaptive 360° Video Streaming: Solutions, Challenges and Opportunities. IEEE Commun. Surv. Tutorials. 2020;22:2801–2838. doi: 10.1109/COMST.2020.3006999. [DOI] [Google Scholar]
  • 96.Brunke L., Greeff M., Hall A.W., Yuan Z., Zhou S., Panerati J., Schoellig A.P. Safe Learning in Robotics: From Learning-Based Control to Safe Reinforcement Learning. Annu. Rev. Control Robot. Auton. Syst. 2022;5:411–444. doi: 10.1146/annurev-control-042920-020211. [DOI] [Google Scholar]
  • 97.Spielberg S., Tulsyan A., Lawrence N.P., Loewen P.D., Bhushan Gopaluni R. Toward self-driving processes: A deep reinforcement learning approach to control. AIChE J. 2019;65 doi: 10.1002/aic.16689. [DOI] [Google Scholar]
  • 98.Liang X., Wang T., Yang L., Xing E. In: Computer Vision – ECCV 2018. Lecture Notes in Computer Science, vol. 11211. Ferrari V., Hebert M., Sminchisescu C., Weiss Y., editors. 2018. CIRL: Controllable Imitative Reinforcement Learning for Vision-Based Self-Driving. (Springer, Cham) [DOI] [Google Scholar]
  • 99.Schulman J., Wolski F., Dhariwal P., Radford A., Klimov O. Proximal Policy Optimization Algorithms. arXiv. 2017 doi: 10.48550/arXiv.1707.06347. Preprint at. [DOI] [Google Scholar]
  • 100.Rosenfeld R. Two decades of statistical language modeling: where do we go from here? Proc. IEEE. 2000;88:1270–1278. doi: 10.1109/5.880083. [DOI] [Google Scholar]
  • 101.Bahdanau D., Cho K., Bengio Y. Neural Machine Translation by Jointly Learning to Align and Translate. arXiv. 2016 doi: 10.48550/arXiv.1409.0473. Preprint at. [DOI] [Google Scholar]
  • 102.Izacard G., Grave E. Leveraging Passage Retrieval with Generative Models for Open Domain Question Answering. arXiv. 2021 doi: 10.48550/arXiv.2007.01282. Preprint at. [DOI] [Google Scholar]
  • 103.Radford A., Narasimhan K., Salimans T., Sutskever I. 2018. Improving Language Understanding by Generative Pre-training; p. 12. [Google Scholar]
  • 104.Radford A., Wu J., Child R., Luan D., Amodei D., Sutskever I. 2019. Language Models Are Unsupervised Multitask Learners; p. 24. [Google Scholar]
  • 105.Ouyang L., Wu J., Jiang X., Almeida D., Wainwright C.L., Mishkin P., Zhang C., Agarwal S., Slama K., Ray A., et al. Training language models to follow instructions with human feedback. arXiv. 2022 doi: 10.48550/arXiv.2203.02155. Preprint at. [DOI] [Google Scholar]
  • 106.Ray P.P. ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope. Internet Things Cyber-Phys. Syst. 2023;3:121–154. doi: 10.1016/j.iotcps.2023.04.003. [DOI] [Google Scholar]
  • 107.Touvron H., Lavril T., Izacard G., Martinet X., Lachaux M.-A., Lacroix T., Rozière B., Goyal N., Hambro E., Azhar F., et al. LLaMA: Open and Efficient Foundation Language Models. arXiv. 2023 doi: 10.48550/arXiv.2302.13971. Preprint at. [DOI] [Google Scholar]
  • 108.Benegas G., Batra S.S., Song Y.S. DNA language models are powerful predictors of genome-wide variant effects. Proc. Natl. Acad. Sci. USA. 2023;120 doi: 10.1073/pnas.2311219120. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 109.Madani A., Krause B., Greene E.R., Subramanian S., Mohr B.P., Holton J.M., Olmos J.L., Xiong C., Sun Z.Z., Socher R., et al. Large language models generate functional protein sequences across diverse families. Nat. Biotechnol. 2023;41:1099–1106. doi: 10.1038/s41587-022-01618-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 110.Liu P., Ren Y., Tao J., Ren Z. GIT-Mol: A Multi-modal Large Language Model for Molecular Science with Graph, Image, and Text. Comput. Biol. Med. 2024;171 doi: 10.1016/j.compbiomed.2024.108073. [DOI] [PubMed] [Google Scholar]
  • 111.Language models can learn complex molecular distributions. Nat. Commun. 2022;13 doi: 10.1038/s41467-022-30839-x. https://www.nature.com/articles/s41467-022-30839-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 112.Bommasani R., Hudson D.A., Adeli E., Altman R., Arora S., von Arx S., Bernstein M.S., Bohg J., Bosselut A., Brunskill E., et al. On the Opportunities and Risks of Foundation Models. arXiv. 2022 doi: 10.48550/arXiv.2108.07258. Preprint at. [DOI] [Google Scholar]
  • 113.Hao Z., Liu S., Zhang Y., Ying C., Feng Y., Su H., Zhu J. Physics-Informed Machine Learning: A Survey on Problems, Methods and Applications. arXiv. 2023 doi: 10.48550/arXiv.2211.08064. Preprint at. [DOI] [Google Scholar]
  • 114.Mao Z., Jagtap A.D., Karniadakis G.E. Physics-informed neural networks for high-speed flows. Comput. Methods Appl. Mech. Eng. 2020;360 [Google Scholar]
  • 115.Lu L., Jin P., Pang G., Zhang Z., Karniadakis G.E. Learning nonlinear operators via DeepONet based on the universal approximation theorem of operators. Nat. Mach. Intell. 2021;3:218–229. doi: 10.1038/s42256-021-00302-5. [DOI] [Google Scholar]
  • 116.Wang S., Wang H., Perdikaris P. Learning the solution operator of parametric partial differential equations with physics-informed DeepONets. Sci. Adv. 2021;7 doi: 10.1126/sciadv.abi8605. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 117.Ma M., Ma T., Guo L.J. 2024. Solving Out-Of-Distribution Challenges in Optical Foundation Models Using Self-Improving Data Augmentation. [Google Scholar]

Articles from iScience are provided here courtesy of Elsevier

RESOURCES