INVESTIGATION OF INVERSE DESIGN OF MULTILAYER THIN -FILMS WITH CONDITIONAL INVERTIBLE NEURAL NETWORKS

2025-05-03 0 0 1.2MB 12 页 10玖币
侵权投诉
INVESTIGATION OF INVERSE DESIGN OF MULTILAYER
THIN-FILMS WITH CONDITIONAL INVERTIBLE NEURAL
NETWORKS
A PREPRINT
Alexander Luce
University Erlangen-Nürnberg
ams OSRAM
Regensburg
Ali Mahdavi
ams OSRAM
Regensburg
Heribert Wankerl
ams OSRAM
Regensburg
Florian Marquardt
University Erlangen-Nürnberg
Max Planck Institute for the Science of Light
Erlangen
October 11, 2022
ABSTRACT
The task of designing optical multilayer thin-films regarding a given target is currently solved using
gradient-based optimization in conjunction with methods that can introduce additional thin-film
layers. Recently, Deep Learning and Reinforcement Learning have been been introduced to the task
of designing thin-films with great success, however a trained network is usually only able to become
proficient for a single target and must be retrained if the optical targets are varied. In this work, we
apply conditional Invertible Neural Networks (cINN) to inversely designing multilayer thin-films
given an optical target. Since the cINN learns the energy landscape of all thin-film configurations
within the training dataset, we show that cINNs can generate a stochastic ensemble of proposals for
thin-film configurations that that are reasonably close to the desired target depending only on random
variables. By refining the proposed configurations further by a local optimization, we show that the
generated thin-films reach the target with significantly greater precision than comparable state-of-the
art approaches. Furthermore, we tested the generative capabilities on samples which are outside the
training data distribution and found that the cINN was able to predict thin-films for out-of-distribution
targets, too. The results suggest that in order to improve the generative design of thin-films, it is
instructive to use established and new machine learning methods in conjunction in order to obtain the
most favorable results.
Keywords Invertible Neural Networks ·optical multilayer thin-films ·latent space
1 Introduction
In optics, being able to develop devices which manipulate light in a desired way is a key aspect for all applications
within the field such as illumination [
1
] or integrated photonics [
2
]. Recent developments in machine learning, deep
learning and inverse design offer new possibilities to engineer such optical and photonic devices [
3
,
4
,
5
,
6
,
7
,
8
].
Nanophotonics in particular benefits from the recent advancements in optimization and design algorithms [
9
,
10
]. For
example the development of meta optics or the design of scattering nano particles was greatly improved by employing
gradient-based inverse design and deep learning [
11
,
12
,
13
]. Multilayer thin-films are another instance of nanophotonic
devices which are employed to fulfill a variety of different functionalities. Application examples are vertical-cavity
surface-emitting lasers [
14
,
15
], anti-reflection coatings [
16
] and wavelength demultiplexers [
17
]. Recently, they were
arXiv:2210.04629v1 [physics.comp-ph] 10 Oct 2022
APREPRINT - OCTOBER 11, 2022
successfully employed to enhance the directionality of a white LED while maintaining the desired color temperature
[
18
]. Designing multilayer thin-films [
19
,
20
,
21
] has been a task in the nanophotonics community for a long time
and sophisticated techniques for the synthesis of thin-films, which exhibit desired optical characteristics have been
developed in open-source or commercially available software [
22
,
23
,
24
,
25
,
26
]. Methods such as the Fourier method
[
27
,
28
] or the needle method [
29
,
28
,
30
,
21
] compute the position inside the thin-film where the introduction of a new
layer is most beneficial. Then the software will continue with a refinement process, often based on a gradient-based
optimization such as the Levenberg-Marquardt algorithm [
31
,
32
], until it reaches a local minimum where it will
then introduce another layer. Although the software will often converge to a satisfying solution with respect to the
given target, the presented solutions often use excessive amounts of layers and the optimization is still limited by the
selected parameters in the beginning of the optimization. The problem of converging to local optima was tackled in
the past by the development of numerous global optimization techniques which have been introduced and tested in
the field of thin-film optimization [
33
,
34
,
35
,
36
,
37
]. Recently, the innovations of machine learning attracted much
interest in the thin-film community and resulted in interesting new ways to create thin films [5, 38]. Particularly, deep
reinforcement learning or Q-learning showed promising results in designing new and efficient multilayer thin-films
while punishing complicated designs, which employ many layers [
39
,
40
] and require targets that are difficult to achieve
with conventional optimization.
In this work we employ so called conditional Invertible Neural Networks (cINNs) [
41
] to directly infer the loss
landscape of all thin-film configurations with a fixed number of layers and material choice. The cINN learns to map the
thin-film configuration to a latent space, conditional on the optical properties, ie. the reflectivity of a thin-film. During
inference, due to the invertibility of the architecture, the cINN maps selected points from the latent space to their most
likely thin-film configurations, conditional on a chosen target. This results in requiring only a single application of the
cINN to obtain the most likely thin-film configuration given an optical target. Additionally, the log-likelihood training
makes the occurrence of mode-collapse [
42
] almost impossible. For thin-films, many different configurations lead to
similar optical properties. For conventional optimization, this leads to the convergence of the optimization to possibly
unfavorable local minima. A cINN circumvents this due to the properties of the latent space - by varying the points
in the latent space, a perfectly trained cINN is able to predict any possible thin-film configuration that satisfies the
desired optical properties. In this work, we investigated how good the generative capabilities of a cINN are for finding
suitable thin-film configurations in a real-world application. We present an optimization algorithm, which is suitable to
improve the thin-film predictions of the cINN. Then, we compared the optimization results of the presented algorithm to
state-of-the-art software. Finally, we discuss the limitations of the approach and give a guideline when the application
of a cINN is advantageous.
2 Normalizing flows and conditional invertible neural networks
Invertible neural networks are closely related to normalizing flows, which were first popularized by Dinh et. al. [
43
]. A
normalizing flow is an architecture that connects two probability distributions by a series of invertible transformations.
The idea is to map a complex probability distribution to a known and simple distribution such as a Gaussian distribution.
This can be used both for density estimation, but also for sampling since points can easily be sampled with a Gaussian
distribution and mapped to the complex distribution via the normalizing flow. The architecture of a normalizing flow is
constructed from the following. Assume two probability distributions,
π
which is known and for which
zπ(z)
holds
and the complex, unknown distribution p. The mapping between both is given by the change-of-variables formula
p(x) = π(z)
det z
x
.(1)
Consider a transformation fwhich maps f(x) = z. Then the change-of-variables formula can be written as
p(x) = π(z)
det f(x)
x
.(2)
The transformation
f
can be given by a series of invertible transformations
f=fKfK1. . . f0
with
x=zK=
f(z0)=(fK. . . f0)(z0)
. Then, the probability density at any intermediate point is given by
pi(xi) = zi=fi(zi1)
.
By rewriting the change-of-variables formula and taking the logarithm one obtains
log (p(x)) = log π(z0)
K
Y
i=1
det fi(zi1)
zi1
1!= log (π(z0))
K
X
i=1
log
det fi(zi1)
zi1
.(3)
To be practical, a key component of any transformation of a normalizing flow is that the Jacobian determinant of
the individual transformations must be easy to compute. A suitable invertible transformation, which is sufficiently
2
APREPRINT - OCTOBER 11, 2022
expressive is the so called RNVP block [
44
]. The input
z
is split into two separate vectors
u1
and
u2
and is processed
with the help of two transformation functions, s2and t2, while the other input u2is kept fixed
fi(u1u2) = {u1exp(s2(u2)) t2(u2)u2}={v1u2}.(4)
,,
and
denote element wise computation while
denotes a concatenation. The inverse of this transformation
is given by
f1
i(v1u2) = {(v1t2(u2)) exp(s2(u2)) u2}.(5)
To invert the entire transformation, no inversion of the transformations
s2
and
t2
is required. Therefore, a neural network
can be used as a transformation to make the normalizing flow expressive. The Jacobian of the transformation is given
by an upper triangular matrix. Therefore, the Jacobian determinant is easy to compute since only the diagonal elements
contribute.
det fi
z =det 1d0
v1
u2diag(exp(s2(u2) + t2(u2))).(6)
Since only one part of the input is transformed by a neural network, the RNVP block is repeated and applied to the yet
untransformed part of the input
u2
with two additional transformation functions
s1
and
t1
. Other transformations, which
can be advantageous depending on the specific application are the GLOW transformation [
45
], which utilizes invertible
1x1 convolutions or the masked autoregressive flow with a generalized implementation of the RNVP block [46].
Utilizing normalizing flows for the task of a generative neural network for the proposition of thin-films requires some
modification of the normalizing flow. Ardizzone et. al. [
41
] propose a so called conditional invertible neural network
(cINN) which extends the change-of-variables formula to conditional probability densities with condition c
p(x|c) = π(z)
det f(x;c)
x
.(7)
By assuming a Gaussian probability distribution for
π
and taking the logarithm, the conditional maximum likelihood
loss function is derived by Ardizzone et. al. for training of a cINN as
LcML =E"kf(x;c)k2
2
2log
det f(x;c)
x #.(8)
The condition
c
can be the result of another neural network
ϕ(c) = s
, which extracts features from the given condition
c
. The features are then passed to the RNVP transformations by appending the features to the input vectors
u1
and
u2
. By jointly training the cINN via the conditional maximum likelihood loss, Ardizzone et. al. showed that the
feature extraction network learns to extract useful information for the invertible transformation. The condition
c
can be
thought of as the target of the cINN. During inference, the cINN transforms gaussian samples to the learned distribution,
conditional on the target
c
. Ardizzone et. al. also provide a Python implementation FrEIA
1
for conditional and regular
invertible neural network based on the Pytorch2Deep Learning library.
3 Application of conditional invertible neural networks to generating multilayer thin-films
Multilayer thin-films are an optical component that consists of a sequence of planar layers with different materials
stacked on top of each other with a varying layer thickness. Light, which irradiates the thin-film, can be transmitted and
reflected at the layer interfaces and absorbed within the layer, as depicted in Figure 2. This interaction of reflection,
transmission and absorption is different for different wavelengths of light and angles of incidence
Θ
. A fast and
convenient way to compute the optical response of a thin-film is to employ the transfer matrix method (TMM) [
47
]. In
a previous work, we developed the Python package TMM-Fast
3
[
48
] which allows to compute the optical response of a
thin-film. The package also implements convenience functionality for thin-film-dataset generation, which is especially
important for machine learning. By changing the layer thicknesses, the behavior of the thin-film can be modified, which
are the parameters pthat are up to optimization.
Finding a thin-film design, which fulfills the optical criteria while employing only a very limited number of layers is a
challenging task since the loss landscape is highly non-convex and high-dimensional. The loss landscape is given by a
mapping of the optical characteristics
M
of the thin-films by a loss function
L(M, λ),Mtarget, λ)) = E
with
1https://github.com/VLL-HD/FrEIA#papers
2https://pytorch.org/
3https://github.com/MLResearchAtOSRAM/tmm_fast
3
摘要:

INVESTIGATIONOFINVERSEDESIGNOFMULTILAYERTHIN-FILMSWITHCONDITIONALINVERTIBLENEURALNETWORKSAPREPRINTAlexanderLuceUniversityErlangen-NürnbergamsOSRAMRegensburgAliMahdaviamsOSRAMRegensburgHeribertWankerlamsOSRAMRegensburgFlorianMarquardtUniversityErlangen-NürnbergMaxPlanckInstitutefortheScienceofLightEr...

展开>> 收起<<
INVESTIGATION OF INVERSE DESIGN OF MULTILAYER THIN -FILMS WITH CONDITIONAL INVERTIBLE NEURAL NETWORKS.pdf

共12页,预览3页

还剩页未读, 继续阅读

声明:本站为文档C2C交易模式,即用户上传的文档直接被用户下载,本站只是中间服务平台,本站所有文档下载所得的收益归上传人(含作者)所有。玖贝云文库仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。若文档所含内容侵犯了您的版权或隐私,请立即通知玖贝云文库,我们立即给予删除!
分类:图书资源 价格:10玖币 属性:12 页 大小:1.2MB 格式:PDF 时间:2025-05-03

开通VIP享超值会员特权

  • 多端同步记录
  • 高速下载文档
  • 免费文档工具
  • 分享文档赚钱
  • 每日登录抽奖
  • 优质衍生服务
/ 12
客服
关注