2021
DOI: 10.1364/oe.426968
|View full text |Cite
|
Sign up to set email alerts
|

Achieving efficient inverse design of low-dimensional heterostructures based on a vigorous scalable multi-task learning network

Abstract: A scalable multi-task learning (SMTL) model is proposed for the efficient inverse design of low-dimensional heterostructures and the prediction of their optical response. Specifically, several types of nanostructures, including single and periodic graphene-Si heterostructures consisting of n×n graphene squares (n=1∼9), 1D periodic graphene ribbons, 2D arrays of graphene squares, pure Si cubes and their periodic array counterparts, are investigated using both traditional finite element method and SMTL network, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6

Relationship

2
4

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 57 publications
0
5
0
Order By: Relevance
“…In general, the designed deep learning models are trained with small datasets, which may lead to poor generalization ability and serious overfitting performance. To cope with the shortcomings, many technical means, such as data augmentation ( 31 ), multi-task learning ( 33 ), transfer learning ( 34 ), and attention mechanism ( 35 ) can be used to achieve the segmentation task. In this section, a transfer learning strategy is applied to solve the shortcomings.…”
Section: Methodsmentioning
confidence: 99%
“…In general, the designed deep learning models are trained with small datasets, which may lead to poor generalization ability and serious overfitting performance. To cope with the shortcomings, many technical means, such as data augmentation ( 31 ), multi-task learning ( 33 ), transfer learning ( 34 ), and attention mechanism ( 35 ) can be used to achieve the segmentation task. In this section, a transfer learning strategy is applied to solve the shortcomings.…”
Section: Methodsmentioning
confidence: 99%
“…Therefore, a normalization method is highly desired to make the above parameters be on the same scale. Thus, the parameter normalization method can be represented as [ 45 ]: where x represents the actual value, x min stands for the minimum value, and δ is the distance of discrete attributes. Notably, the dataset contains only reflectance spectra with non-negative incident angles are included in the dataset due to the symmetry feature, with the incident angle resolution being 1°.…”
Section: Methodsmentioning
confidence: 99%
“…Here, “optimization” means reducing the discrepancies between the current spectrum and the target spectrum to obtain more precise used parameters by changing the incident angle and structural variable, whose adjustment process is defined as “action”. Moreover, “greedy” indicates that the model selects actions that better reduce the spectral differences, characterized by the usage of mean absolute percentage error (MAPE) [ 45 ]: …”
Section: Methodsmentioning
confidence: 99%
See 1 more Smart Citation
“…[41] Additionally, we abstract the DOFCN as a composite vector convolution (CVC) operator and build CVC neural networks (CVCNNs). Subsequently, we select signal modulation format identification (MFI) and optical metamaterial structure inverse design (SID) [42,43] tasks to test the classification and regression capability of CVCNNs, respectively. The ablation experiments reveal that CVCNNs have improved performance than that of the Baseline (32-bit CNNs) reported in refs.…”
Section: Introductionmentioning
confidence: 99%