You are currently viewing a new version of our website. To view the old version click .
Materials
  • Article
  • Open Access

23 October 2025

Deep Learning-Based Inverse Design of Stochastic-Topology Metamaterials for Radar Cross Section Reduction

,
,
,
and
1
School of Electronic Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2
National Institute of Defense Technology Innovation, Academy of Military Sciences PLA China, Beijing 100071, China
*
Authors to whom correspondence should be addressed.
This article belongs to the Section Materials Simulation and Design

Abstract

Electromagnetic (EM) metamaterials have a wide range of applications due to their unique properties, but their design is often based on specific topological structures, which come with certain limitations. Designing with stochastic topologies can provide more diverse EM properties. However, this requires experienced designers to search and optimise in a vast design space, which is time-consuming and requires substantial computational resources. In this paper, we employ a deep learning network agent model to replace time-consuming full-wave simulations and quickly establish the mapping relationship between the metamaterial structure and its electromagnetic response. The proposed framework integrates a Convolutional Block Attention Module-enhanced Variational Autoencoder (CBAM-VAE) with a Transformer-based predictor. Incorporating CBAM into the VAE architecture significantly enhances the model’s capacity to extract and reconstruct critical structural features of metamaterials. The Transformer predictor utilises an encoder-only configuration that leverages the sequential data characteristics, enabling accurate prediction of electromagnetic responses from latent variables while significantly enhancing computational efficiency. The dataset is randomly generated based on the filling rate of unit cells, requiring only a small fraction of samples compared to the full design space for training. We employ the trained model for the inverse design of metamaterials, enabling the rapid generation of two cells for 1-bit coding metamaterials. Compared to a similarly sized metallic plate, the designed coding metamaterial radar cross-section (RCS) reduces by over 10 dB from 6 to 18 GHz. Simulation and experimental measurement results validate the reliability of this design approach, providing a novel perspective for the design of EM metamaterials.

1. Introduction

Electromagnetic (EM) metamaterials can realise precise control over the absorption and scattering of electromagnetic waves by designing effective structural units. They find widespread applications in energy conversion, sensors, and cloaking technology [,,,]. In recent years, coding metamaterials have attracted much attention because they can realise more precise and flexible control of EM waves by designing specific coding sequences [,]. Traditional design methods typically involve unit structure design based on design objectives and coding schemes, parameter scanning using electromagnetic simulation software, and optimisation and verification. However, this design approach is highly complex, requiring significant time and computational resources []. Moreover, it heavily relies on the designer’s expertise, exhibits limited topological structure diversity, and imposes restrictions on the design space, making it potentially inadequate for specific design requirements.
To enhance design freedom, stochastic-topology-based metamaterial design methods have been employed []. By algorithmically generating rapidly stochastic topological structures, diverse and unique metamaterial configurations can be created, exhibiting distinct EM properties. However, optimising the structure to meet the design objectives becomes more time-consuming as design freedom increases. Many researchers have utilised evolutionary algorithms, such as particle swarm optimisation (PSO) and genetic algorithms (GA), to search for global optimal solutions within the design space through heuristic updating rules [,]. These algorithms are easy to implement but still require iterations with simulation software, and the computational cost grows exponentially as the dimensionality of the design variables increases.
Currently, due to the superior performance of deep learning, several scholars have integrated deep learning with metamaterials design, proposing various deep learning-based inverse design methods [,,]. This category of research primarily encompasses two aspects: EM structure generation and EM response prediction. In EM structure generation models, generative adversarial networks (GAN) produce metamaterial unit structures with high visual fidelity through adversarial training, where the generator synthesises complex topologies and the discriminator evaluates structural rationality [,,]. However, this necessitates a substantial amount of training data, particularly for structures with high design freedom such as stochastic-topology metamaterials. The variational autoencoder (VAE) comprises an encoder and a decoder, featuring a simple architecture that facilitates efficient training [,]. This model has gained widespread adoption in the design of EM structures in recent years [,]. The encoder compresses structural information into a low-dimensional latent space, with sampled vectors subsequently reconstructed or generated by the decoder. Among numerous related studies, further research is required to explore how to improve reconstruction quality and enhance the model’s feature extraction capability. For EM response prediction, latent space variables serve as inputs, while spectral sequences constitute the outputs []. The Transformer model, as a core architecture for sequence modelling, has extensive applications in natural language processing, time-series forecasting, and cross-modal learning []. The original Transformer employs an encoder-decoder architecture that treats spectral sequences as temporally dependent outputs, which easily leads to error accumulation. In contrast, the encoder-only configuration establishes direct mappings between latent variables and spectral sequences through global self-attention mechanisms, thereby eliminating redundant decoder modules to significantly boost computational efficiency [].
In summary, this work proposes a neural network-based assisted inverse design algorithm for rapid design of stochastic topological metamaterials [,]. The novelty of our approach is its integrated deep learning framework, which provides an efficient and scalable solution for the inverse design of stochastic-topology metamaterials by synergizing a CBAM-enhanced VAE with an encoder-only Transformer. Specifically, this methodology integrates a convolutional block attention module (CBAM) [] into the VAE framework to enhance feature extraction capability, while employing an encoder-only architecture for predicting electromagnetic response amplitude and phase of stochastically configured materials. This approach maintains modelling capacity while significantly improving computational efficiency. Furthermore, we implement a multi-task weighted loss function to balance structural reconstruction and response prediction. Ultimately, the proposed network is used to replace simulation software for the design of 1-bit coding metamaterials, achieving a Radar Cross Section (RCS) reduction of more than 10 dB across the 6–18 GHz frequency range.

2. Theories and Methods

2.1. Stochastic Topology Metamaterials

In this work, as shown in Figure 1a, the stochastic topology metamaterial unit consists of four layers: dielectric-impedance-dielectric-metal, with a unit cell period P = 8 mm. The upper and lower dielectric layers are made of F4BTM300 with a relative permittivity of εr = 3.0 and a loss tangent of tan θ = 0.001, both having a thickness of t1 = t2 = 3.5 mm. The impedance layer, implemented using a nickel-chromium (NiCr) thin-film (Ticer TCR®, Ticer Technologies) with a sheet resistance of 100 ohm/sq, follows the p4m rule [], resulting in an axisymmetric structure as illustrated in Figure 1b. Subsequently, the impedance layer structure is encoded using “1” and “0” to represent the presence or absence of the impedance film, following the coding rules depicted in Figure 1c. The entire structure is divided into N × N small cells. Since the generated image exhibits an axisymmetric structure, a randomly generated coding matrix with a size of N/2 × N/2 is flipped and superimposed, followed by two symmetries, to generate the complete structure. The bottom conductor is composed of a copper foil with a thickness of tpec = 0.017 mm.
Figure 1. Schematic of the metamaterial unit cell and its coding rules: (a) the layered structure of the unit cell, from top to bottom: a dielectric layer, an impedance layer, a second dielectric layer, and a metallic plate; (b) the topology of the impedance layer; (c) the structural coding rules.

2.2. Deep Learning Approach

To achieve topology reconstruction and EM response prediction, we architected a multi-task deep learning network. As illustrated in Figure 2, the framework integrates a CBAM-enhanced VAE module and a Transformer predictor, respectively performing metamaterial structural reconstruction and electromagnetic response prediction.
Figure 2. Schematic of the multi-task deep learning network. (a) Overall architecture of CBAM-VAE-Transformer network; (b) Structure of CBAM attention mechanism; (c) Configuration of Transformer-based predictor.

2.2.1. CBAM-VAE

The VAE is a deep learning model that integrates Bayesian networks, primarily comprising an encoder and a decoder, as depicted in Figure 2a []. Here, x denotes the input stochastic topological structure, z represents the latent variable, and x′ signifies the reconstructed or generated stochastic topological structure. The encoder first maps the high-dimensional input x to an approximate posterior distribution q ϕ ( z | x ) over the latent variables z, assumed to follow a normal distribution,
q ϕ ( z | x ) = N ( z ; μ , σ 2 I )
where μ and σ 2 denote the mean and variance of the normal distribution, I is the identity matrix, and ϕ represents variational parameters. Since the relationship between latent variable z and parameters ϕ is non-deterministic, directly sampling z from the encoder is non-differentiable, preventing gradient backpropagation. To resolve this, a stochastic sample ε is introduced, enabling differentiable sampling of latent variable z via the reparameterization trick,
z = μ + σ ε
where ε follows a standard normal distribution N ( 0 , I ) . Finally, the decoder reconstructs the latent variable z into x′.
To enhance the feature extraction and reconstruction capabilities of the VAE model, CBAM is introduced. As a lightweight attention module that combines channel and spatial dimensions, CBAM comprises two submodules: the channel attention module (CAM) and the spatial attention module (SAM), as shown in Figure 2b. For an intermediate feature map, CBAM sequentially infers attention maps along two separate dimensions (channel and spatial), then multiplies these maps with the input feature map for adaptive feature refinement.
The CAM aggregates spatial information through concurrent global average pooling (AvgPool) and max pooling (MaxPool), processed via a shared multilayer perceptron (MLP) to generate a 1D channel attention vector. This operation compresses spatial dimensions without altering channel depth, focusing on semantically meaningful information within the input feature map. The channel attention mechanism can be expressed as:
M c ( F ) = δ ( M L P ( A v g P o o l ( F ) ) + M L P ( M a x P o o l ( F ) ) )   = δ ( W 1 m ( W 0 m ( F Avg   c ) ) + W 1 m ( W 0 m ( F Max   c ) ) )
where M c ( F ) denotes the CAM output, and δ represents the sigmoid operation. W 1 m and W 0 m indicate weights of the two-layer MLP.
Subsequently, the SAM processes the channel-refined feature map through separate average pooling (AvgPool) and maximum pooling (MaxPool) operations along the channel axis. The resulting outputs are concatenated and processed by a convolution layer to generate a 2D spatial attention map. These element-wise multiplied attention maps are applied to the input feature map for adaptive feature refinement. This module preserves spatial dimensionality while compressing channel depth, focusing on target positional information. The spatial attention mechanism is formulated as:
M s ( F ) = δ ( f 7 × 7 [ A v g P o o l ( F ) ; M a x P o o l ( F ) ] )   = δ ( f 7 × 7 ( [ F Avg   s ; F Max   s ] ) )
where M s ( F ) denotes the SAM output, and f 7 × 7 indicates the convolution kernel size. As CBAM is a lightweight general-purpose module, it incurs negligible overhead and can be directly integrated into the VAE architecture.
The CBAM-VAE model training is performed by maximising the evidence lower bound (ELBO), equivalent to minimising the following loss function,
L V A E = L r e c o n + L K L
where L r e c o n denotes the reconstruction error, quantifying the accuracy of decoder output x′; and L K L represents the Kullback-Leibler divergence loss, constraining the approximate posterior distribution q ϕ ( z | x ) N ( μ , σ 2 I ) to converge toward the isotropic Gaussian prior p ( z ) N ( 0 , I ) . The synergistic optimisation of reconstruction and KL terms enables the model to learn a well-structured latent space while maintaining effective data reconstruction capability.

2.2.2. Transformer Predictor

In the proposed architecture, the transformer predictor module employs an encoder-only configuration to map latent variables z to multi-frequency EM response predictions. As illustrated in Figure 2c, the structure comprises an input layer, a transformer encoder, and an output layer. To satisfy the transformer encoder’s input-output sequence length equivalence, the 16-dimensional latent variable z is first projected into a 64-dimensional feature space via linear projection, followed by sequence replication to construct an 81-length frequency-position sequence. Learnable positional encodings are incorporated into the input layer and added to the expanded sequence, serving as input to the transformer encoder.
The multi-layer transformer encoder forms the core component, extracting features from latent variable sequences while establishing global dependencies among frequency points. Each encoder layer consists of two sub-layers: a multi-head self-attention (MHSA) layer and a feed-forward network (FFN) layer. Both sub-layers employ residual connections and layer normalisation to stabilise training and prevent gradients from vanishing or exploding. The MHSA mechanism comprises multiple parallel attention heads where input sequences undergo linear transformations to generate query (Q), key (K), and value (V) vectors. Attention outputs from all heads are concatenated and fused through linear projection, computed as follows,
A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d k ) V M u l t i H e a d ( Q , K , V ) = C o n c a t ( h e a d 1 , , h e a d h ) w h e r e       h e a d i = A t t e n t i o n ( Q i , K i , V i )
where d k denotes the dimensionality of K, Softmax serves as the activation function to normalize attention scores into a probability distribution [], and Concat is the concatenation operation that combines outputs from multiple attention heads []. The FFN employs dual linear transformations with ReLU activation to enhance model representational capacity, formulated as,
F F N ( x ) = max ( 0 , W 1 l + b 1 ) W 2 l + b 2
where W 1 l and W 2 l are weight matrices, b 1 and b 2 denote bias vectors, and max ( , ) represents the ReLU activation function. Finally, the output layer transforms features at each frequency point into the electromagnetic response domain via linear transformation, while applying a sigmoid activation to constrain outputs to [0,1],
S ( x ) = 1 1 + e x

2.3. Data Gathering and Model Training

In this paper, we employed PyCharm-CST (CST STUDIO SUITE 2023) [] to conduct joint simulations on metamaterials and obtain EM response curves, thereby enhancing the integration of subsequent deep learning networks. Initially, stochastic-topology structures were generated according to the p4m rule and recorded in a coding form []. Subsequently, the generated GDS files were imported into CST for modelling. The simulation utilised a time-domain solver to solve within the 0 to 20 GHz frequency range. The process was fully automated, generating a dataset of 70,000 samples. The full-wave simulation for each unit cell required approximately 1 min, amounting to a total computational time of around 70,000 min. It is specifically noted that our dataset was randomly generated according to unit cell fill ratios to characterise the stochastic topology design space comprehensively [].
Next, the generated data underwent preprocessing. To comprehensively describe the information of the stochastic-topology structures, the randomly generated 40 × 40 encoding matrix was expanded to an 80 × 80 matrix, serving as the input for the network structure. Additionally, due to the excessive length of the data exported from CST, we performed sampling to reduce its dimensionality. The sampling interval was set to 25, resulting in a data length of 81 after sampling. Moreover, we addressed the discontinuity of the phase curves by mapping them to the x and y axes of the unit circle, as demonstrated in Figure 3. Finally, the amplitude and phase of the response curves were normalised and utilised as outputs for the deep learning network.
Figure 3. Projecting discontinuous phases on the x− and y− axes of the unit circle.
The designed multi-task deep learning network, illustrated in Figure 2a, comprises three distinct objectives: structure reconstruction, KL regularisation, and electromagnetic response prediction. Given their varying learning complexities and convergence rates, a weighted loss function L is employed,
L = λ 1 L r e c o n + λ 2 L K L + λ 3 L p r e L r e c o n = 1 N i = 1 N ( x i l o g ( x i ) + ( 1 x i ) l o g ( 1 x i ) ) L K L = 1 2 N i = 1 N ( σ i 2 + μ i 2 l o g ( σ i 2 ) 1 ) L p r e = 1 N i = 1 N ( y i y i )
where L r e c o n denotes the metamaterial structural reconstruction loss, L K L represents the Kullback-Leibler divergence loss, and L p r e signifies the EM response prediction loss. Here, N indicates the sample size, while λ 1 , 2 , 3 are hyperparameters that coordinate multi-task optimisation and balance learning priorities.
Phase-dependent weighting dynamically adjusts learning focus: higher initial weights on L r e c o n and L p r e capture foundational features, while progressively increasing L K L regularisation in later stages regularises the latent space. This strategy prevents gradient dominance by any single loss term that would degrade other tasks, ultimately achieving a balance between reconstruction fidelity, prediction accuracy, and latent space organisation.
The dataset was randomly split once into a training set (80%) for model development and a held-out test set (20%) for final evaluation. The network was configured with the following parameters: batch_size = 128 and epoch = 120. All computational work, including dataset construction and network training, was performed on a ThinkStation workstation (Windows 11) featuring an NVIDIA Quadro P6000, two Intel Xeon Gold 6154 @ 3.0 GHz CPUs, and 512 GB RAM.

3. Result and Discussion

3.1. Deep Learning Network

Upon completing training, the developed multi-task deep learning network performs two functions: (1) reconstruction or generation of EM structures, and (2) EM response prediction. To evaluate the trained model’s performance, reconstruction accuracy and mean squared error (MSE) are adopted as quantitative metrics,
r e c o n _ a c c = 1 N i = 1 N I ( x i = x i ) p r e _ m s e = 1 M j = 1 M ( y j y j ) 2
where N denotes the total pixel count across all samples, x i and x i represent the i-th pixel of the binarised ground-truth and reconstructed images, respectively, and I is the indicator function equaling 1 when the condition holds and zero otherwise. M signifies the aggregate frequency points across samples, with y i and y i indicating the j-th normalised frequency point value of true and predicted EM responses.
Testing demonstrates 98% reconstruction accuracy and a prediction error of 1 × 10−4 on the test set. Two randomly selected samples, visualised in Figure 4, confirm the precise structural generation alongside accurate amplitude and phase response prediction.
Figure 4. Examples of (a) Structural reconstruction and (b) EM response prediction.

3.2. Inverse Design of 1-Bit Coding Metamaterials

The trained deep learning network enables the efficient generation of stochastic-topology metamaterials that meet specified EM response requirements. As detailed in Figure 5a, the workflow consists of three stages: initialisation, optimisation, and structure generation. During initialisation, the target EM response amplitude and phase y t a r g e t is defined while the latent variable z 0 is randomly sampled in latent space. Gradient descent is performed using the Adam optimiser with learning rate scheduling [].
Figure 5. Flowchart and design example of metamaterial inverse design based on a deep learning network. (a) Inverse design flowchart; (b) Generated structure, optimisation loss, and EM response comparison.
In the iterative optimisation phase:
(1)
The transformer predictor predicts EM responses for current latent variables y;
(2)
The MSE between predictions and targets is computed;
(3)
Optimal latent vector z b e s t is recorded when new minimum MSE occurs;
(4)
Latent vectors z n e w are updated via backpropagation.
Iteration terminates when either the maximum number of iterations (500) is reached and the MSE falls below an acceptable threshold. The optimal z b e s t then generates the final binarised structure x t a r g e t , with its real EM response subsequently calculated.
Figure 5b demonstrates inverse design results for a stochastic specimen. The loss function decreased rapidly during optimization: 1 × 10−3 at iteration 25 and 1 × 10−4 at iteration 132, converging to 3 × 10−5. The entire inverse design process was completed in only 3.2 s on the ThinkStation workstation. This total time includes loading the pre-trained model and executing 500 optimization iterations, representing a speed improvement of several orders of magnitude compared to traditional EM simulation-based methods that would require hundreds of minutes for a comparable design task. A comprehensive assessment must consider the total time investment, including the initial overhead of our deep learning approach. The initial investment requires approximately 1167 h for dataset generation and 1 h for model training. Although substantial, this initial cost is amortized over subsequent designs. A conventional design process relying on 500 sequential full-wave simulations would require approximately 500 min. Analysis indicates that a break-even point is reached when designing more than approximately 140 units. Beyond this point, our method achieves a lower total time investment, with the efficiency advantage becoming substantial for large-scale design tasks.
As an application, we designed a 1-bit coding metamaterial based on stochastic topological structures, where two metamaterial units exhibited a phase difference of 180°. The designed metamaterial units are arranged in a 6 × 6 periodic pattern to form two supercells, labelled as “0” and “1”, as shown in Figure 6. The overall metamaterial consists of an 8 × 8 array of supercells arranged in a chessboard pattern, with dimensions of 384 mm × 384 mm.
Figure 6. Full model of chessboard metamaterial and its supercells. (a) element “0”; (b) chessboard metamaterial; (c) element “1”.
Figure 7 illustrates the simulation results of the monostatic RCS and RCS reduction for the designed coding metamaterial compared to a metallic plate of the same size. The coding metasurface achieved a significant RCS reduction of more than 10 dB within the frequency range of 6–18 GHz (100% relative bandwidth). At 8.5 GHz and 16 GHz, the maximum RCS reduction values are obtained, reaching 32.87 dB and 33.05 dB, respectively. Furthermore, Figure 8 presents the 2D and 3D bistatic RCS patterns at 8.5 GHz and 16 GHz. From Figure 8a,c, we can observe that the incident plane wave is scattered into four main beams: φ = 45°, 135°, 225° and 315°. This scattering behaviour effectively suppresses RCS. The 2D scattering plots of coding metamaterials at 8.5 and 16 GHz are given in Figure 8b,d. The scattered field is effectively suppressed in all directions compared to the PEC. This is attributed to the coding metamaterial design, which disperses the incident electromagnetic wave energy in all directions.
Figure 7. Comparison of RCS reduction simulation results between 1−bit coding metamaterials and metallic plate.
Figure 8. 3D bistatic scattered fields at (a) 8.5 and (b) 16 GHz under normal incidence for the 1-bit coding metamaterial with chessboard structure; Comparison of the bistatic scattered fields along the XZ/YZ planes at (c) 8.5 and (d) 16 GHz.
To validate the proposed methodology against full-wave simulation results, we fabricated and experimentally characterised the chessboard metamaterial prototype illustrated in Figure 9. The manufactured specimen (384 mm × 384 mm × 7 mm) employs dual-layer F4BM300 substrates ( ε r = 3.0 ) with 0.017 mm copper cladding. Figure 9 details the measurement configuration in a compact antenna test range (CATR) chamber, where the prototype was mounted on radar-transparent foam supports. The low-scattering metamaterial was designed for operation in the 6-18 GHz frequency range. However, due to equipment limitations, there was no standard gain horn antenna that directly covers this entire band. Therefore, horn antennas for different frequency segments were used in sections for testing. The frequency bands of the test horn antennas are 5.8–8.2 GHz, 8.2–12.4 GHz, and 12.4–18 GHz, respectively. Each antenna replacement necessitated recalibration and background subtraction. The final full-band response was obtained by stitching the data from these antennas, specifically by averaging the measured values in the overlapping frequency regions (i.e., around 8.2 GHz and 12.4 GHz). Owing to the symmetric configuration of the metamaterial, the structure is polarization-insensitive. Monostatic RCS measurements were performed under co-polarized (HH) excitation at normal incidence.
Figure 9. Fabricated chessboard Metamaterial and compact antenna test range (CATR) test environment. (a) Fabricated metasurface prototype; (b) Target mounted on radar-transparent foam supports; (c) Reflector; (d) Transmit/receive horn antennas.
Under HH-polarised wave normal incidence, the comparison between the measured and simulated monostatic RCS results of the low-scattering checkerboard metamaterial as a function of frequency is shown in Figure 10. The designed metamaterial achieves over 10 dB RCS reduction across the entire 6–18 GHz band compared to reference metal plates of equivalent thickness. The acceptable errors may arise from metamaterial fabrication errors or angular deviations between the metamaterial and the reflecting surface.
Figure 10. Comparison of simulation and measurement results for monostatic RCS (a) and RCS reduction (b).

3.3. Discussion

To further validate the superiority and practicality of the proposed methodology, comparative analyses were conducted against prior works in Table 1 and Table 2 focusing on network architecture and RCS reduction performance. Our network enhances the baseline VAE by integrating CBAM, significantly improving feature extraction capability and structural reconstruction fidelity—achieving higher reconstruction accuracy than conventional autoencoder (AE). The encoder-only Transformer predictor attains lower mean squared error (MSE) for electromagnetic response prediction compared to CNN-based models. Crucially, the proposed method exhibits the highest total permutations, yet it requires only a modest increase in training samples while simultaneously predicting the amplitude and phase of EM response. As observed from Table 2, this work employs stochastic topology, offering higher degrees of freedom, and the designed coding metamaterial achieves a maximum fractional bandwidth of 100% in RCS reduction. This finding validates the effectiveness of the proposed method.
Table 1. Metamaterials Design Methods.
Table 2. Metamaterials for RCS Reduction.

4. Conclusions

This paper proposes a deep learning network-based rapid design methodology for stochastic-topology metamaterials. The VAE model maps high-dimensional structural information to low-dimensional latent variables, enabling the reconstruction and generation of metamaterials. Incorporating CBAM modules enhances the feature extraction capability, achieving a 98% reconstruction accuracy. An encoder-only Transformer predictor accurately forecasts electromagnetic responses (amplitude and phase) from latent variables with a prediction error of 1 × 10−4.
The trained network facilitates inverse metamaterial design, converging to a response error of 3 × 10−5 within 500 iterations while consuming only 3.2 s. For demonstration, 1-bit coding metasurfaces were designed, rapidly yielding two unit cells with a 180° phase difference, which achieve an RCS reduction of over −10 dB in the 6–18 GHz range. Simulation and experimental measurements jointly validate the effectiveness of this approach.
Collectively, our framework replaces computationally intensive full-wave simulations in conventional workflows, enabling efficient inverse design and establishing a promising framework for performance-driven metamaterial synthesis. Despite this success, the model’s generalizability beyond stochastic topologies presents a limitation and an opportunity for future work. Looking ahead, we will focus on multi-objective optimization and active learning to further reduce data dependencies, extending this efficient design approach to a broader range of electromagnetic and multi-physics applications.

Author Contributions

Conceptualization, C.Z. (Chao Zhang) and S.G.; methodology, C.Z. (Chao Zhang); validation, C.Z. (Chao Zhang), C.Z. (Chunrong Zou) and Y.Z.; resources, Y.Z. and T.S.; data curation, C.Z. (Chao Zhang) and S.G.; writing—original draft preparation, C.Z. (Chao Zhang); writing—review and editing, C.Z. (Chao Zhang) and S.G.; supervision, Y.Z. and T.S.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62471120 and in part by Sichuan Science and Technology Program under Grant 2024NSFSC0440.

Data Availability Statement

No new database has been created, please contact the author if required.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, H.L.; Zhang, Y.K.; Zhang, T.Y.; Ma, H.F.; Cui, T.J. Broadband and programmable amplitude-phase-joint-coding information metasurface. ACS Appl. Mater. Interfaces 2022, 14, 29431–29440. [Google Scholar] [CrossRef]
  2. Xue, J.; Jiang, W.; Gong, S. Chessboard AMC surface based on quasi-fractal structure for wideband RCS reduction. IEEE Antennas Wirel. Propag. Lett. 2017, 17, 201–204. [Google Scholar] [CrossRef]
  3. Sun, J.; Zhu, L.; Hao, Z.; Kuai, X.; Cao, Q. Design and application of the switchable electric/magnetic conductor antenna. IEEE Trans. Antennas Propag. 2023, 71, 7102–7112. [Google Scholar] [CrossRef]
  4. Phon, R.; Lee, M.; Lor, C.; Lim, S. Multifunctional reflective metasurface to independently and simultaneously control amplitude and phase with frequency tunability. Adv. Opt. Mater. 2023, 11, 2202943. [Google Scholar] [CrossRef]
  5. Cui, T.J.; Qi, M.Q.; Wan, X.; Zhao, J.; Cheng, Q. Coding metamaterials, digital metamaterials and programmable metamaterials. Light. Sci. Appl. 2014, 3, e218. [Google Scholar] [CrossRef]
  6. Liu, C.; Ma, Q.; Luo, Z.J.; Hong, Q.R.; Xiao, Q.; Zhang, H.C.; Miao, L.; Yu, W.M.; Cheng, Q.; Li, L.; et al. A programmable diffractive deep neural network based on a digital-coding metasurface array. Nat. Electron. 2022, 5, 113–122. [Google Scholar] [CrossRef]
  7. Li, L.; Jun Cui, T.; Ji, W.; Liu, S.; Ding, J.; Wan, X.; Bo Li, Y.; Jiang, M.; Qiu, C.W.; Zhang, S. Electromagnetic reprogrammable coding-metasurface holograms. Nat. Commun. 2017, 8, 197. [Google Scholar] [CrossRef]
  8. Xi, Z.; Lu, X.; Shen, T.; Zou, C.; Chen, L.; Guo, S. Research on Design Method of Multilayer Metamaterials Based on Stochastic Topology. Materials 2023, 16, 5229. [Google Scholar] [CrossRef]
  9. Gu, P.; Cao, Z.; He, Z.; Ding, D. Design of ultrawideband RCS reduction metasurface using space mapping and phase cancellation. IEEE Antennas Wirel. Propag. Lett. 2023, 22, 1386–1390. [Google Scholar] [CrossRef]
  10. Hao, H.; Du, S.; Zhang, T. Small-size broadband coding metasurface for RCS reduction based on particle swarm optimization algorithm. Prog. Electromagn. Res. M 2019, 81, 97–105. [Google Scholar] [CrossRef]
  11. Wang, P.; Wei, Z.; Wu, T.; Jiang, W.; Hong, T.; Pedersen, G.F.; Shen, M. A machine learning framework for the design of STCDME structures in RIS applications. IEEE Trans. Microw. Theory Tech. 2023, 72, 1467–1479. [Google Scholar] [CrossRef]
  12. Naseri, P.; Hum, S.V. A generative machine learning-based approach for inverse design of multilayer metasurfaces. IEEE Trans. Antennas Propag. 2021, 69, 5725–5739. [Google Scholar] [CrossRef]
  13. Niu, C.; Phaneuf, M.; Qiu, T.; Mojabi, P. A deep learning-based approach to design metasurfaces from desired far-field specifications. IEEE Open J. Antennas Propag. 2023, 4, 641–653. [Google Scholar] [CrossRef]
  14. Gu, Z.; Li, D.; Wu, Y.; Fan, Y.; Yu, C.; Chen, H.; Li, E.P. A solution to the dilemma for FSS inverse design using generative models. IEEE Trans. Antennas Propag. 2023, 71, 5100–5109. [Google Scholar] [CrossRef]
  15. Jiang, W.; Zhu, Y.; Yin, G.; Lu, H.; Xie, L.; Yin, M. Dispersion relation prediction and structure inverse design of elastic metamaterials via deep learning. Mater. Today Phys. 2022, 22, 100616. [Google Scholar] [CrossRef]
  16. Huang, Y.; Feng, N.; Cai, Y. Artificial intelligence-generated terahertz multi-resonant metasurfaces via improved transformer and CGAN neural networks. J. Light. Technol. 2023, 42, 1518–1525. [Google Scholar] [CrossRef]
  17. Lin, H.; Hou, J.; Wang, Y.; Tang, R.; Shi, X.; Tian, Y.; Xu, W. Machine-learning-assisted inverse design of scattering enhanced metasurface. Opt. Express 2022, 30, 3076–3088. [Google Scholar] [CrossRef]
  18. Wang, P.; Li, Z.; Luo, C.; Wei, Z.; Wu, T.; Jiang, W.; Hong, T.; Parchin, N.O.; Pedersen, G.F.; Shen, M. Preprocessing-based fast design of multiple EM structures with one deep neural network. IEEE Trans. Antennas Propag. 2024, 72, 4298–4310. [Google Scholar] [CrossRef]
  19. Qu, M.; Chen, J.; Su, J.; Gu, S.; Li, Z. Design of metasurface absorber based on improved deep learning network. IEEE Trans. Magn. 2023, 59, 1–6. [Google Scholar] [CrossRef]
  20. You, X.C.; Lin, F.H. Inverse design of reflective metasurface antennas using deep learning from small-scale statistically random pico-cells. Microw. Opt. Technol. Lett. 2024, 66, e34068. [Google Scholar] [CrossRef]
  21. Wang, J.; Fan, C.; Liao, Y.; Zhou, L. Pattern-to-absorption prediction for multilayered metamaterial absorber based on deep learning. IEEE Microw. Wirel. Technol. Lett. 2024, 34, 463–466. [Google Scholar] [CrossRef]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  23. Niu, C.; Kelly, M.; Mojabi, P. An encoder-only transformer to generate power patterns from far-field performance criteria. In Proceedings of the 2022 16th European Conference on Antennas and Propagation (EuCAP), Madrid, Spain, 27 March–1 April 2022; pp. 1–4. [Google Scholar]
  24. Ma, C.; Wang, Z.; Zhang, H.; Yang, F.; Chen, J.; Ren, Q.; Ma, Y.; Wang, N. Inverse design of electromagnetic metamaterials: From iterative to deep learning-based methods. J. Micromech. Microeng. 2024, 34, 053001. [Google Scholar] [CrossRef]
  25. Chen, Q.; Zhang, L.; Ma, H.; Li, D.; Li, Y.; Liu, E.X.; Li, E.P. Deep Neural Network-Based Surrogate-Assisted Inverse Optimization for High-Speed Interconnects. IEEE Trans. Electromagn. Compat. 2024, 66, 2019–2026. [Google Scholar] [CrossRef]
  26. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  27. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  28. Nielsen, M.A. Neural Networks and Deep Learning; Determination Press: San Francisco, CA, USA, 2015; Volume 25. [Google Scholar]
  29. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  30. Studio, M. CST-computer simulation technology. Bad Nuheimer Str. 2008, 19, 64289. [Google Scholar]
  31. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  32. Hossain, M.B.; Faruque, M.R.I.; Abdullah, S.; Islam, M.T. Optimization-Driven design of a 2-bit coding based metasurface absorber for enhanced EMI shielding and RCS reduction. Ain Shams Eng. J. 2024, 15, 102674. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.