B-OPC_0105
B-OPC_0105
January 5, 2025
Abstract
This paper introduces B-OPC, a novel mask optimization framework
based on Binary Neural Networks (BNNs), designed to enhance compu-
tational efficiency while maintaining high performance in semiconductor
lithography. The proposed approach leverages the inherent binary nature
of mask patterns and design targets, enabling significant reductions in
memory usage and computational overhead. Experimental results demon-
strate that the trained model effectively compensates for diffraction effects
on specific design patterns. For 10 ICCAD 2013 contest design patterns,
B-OPC achieves competitive performance compared to traditional meth-
ods such as Neural-ILT and PGAN-OPC, while substantially reducing
training time and computational resource requirements. The framework
introduces a new direction for deep learning-based mask optimization, of-
fering a scalable and efficient solution for large-scale semiconductor man-
ufacturing.
1 Introduction
Lithography is a critical process in semiconductor manufacturing, where circuit pat-
terns are transferred onto wafers by projecting light through a photomask. As tran-
sistor sizes continue to shrink, optical effects such as diffraction increasingly distort
the projected patterns, making mask optimization essential for achieving accurate pat-
tern reproduction. Optical Proximity Correction (OPC) addresses these distortions
by modifying the mask geometry to minimize deviations between the intended and
1
printed patterns. Traditional OPC methods include model-based approaches [1, 2],
which iteratively simulate and refine patterns, and Inverse Lithography Techniques
(ILT) [3, 4], which solve an inverse problem to derive optimal mask designs. While
these methods yield high-quality results, their high computational cost and runtime
pose significant challenges to scalability, prompting ongoing efforts to improve effi-
ciency and practicality.
In recent years, deep learning methods have emerged as powerful tools for reducing
runtime and enhancing the printability of optimized masks [5, 6, 7, 8, 9, 10]. Unlike
ILT, which relies on iterative gradient descent to minimize loss and refine mask de-
signs, deep neural network (DNN)-based approaches can generate optimized masks
in a single forward pass. This eliminates the need for iterative approximations [11],
significantly improving computational efficiency. Moreover, most DNN-based meth-
ods produce near-optimal mask designs, requiring only minimal refinement through
conventional mask optimization techniques to achieve final results.
Recent studies have integrated various DNN frameworks with traditional OPC and
ILT to further enhance computational efficiency. Yang et al. [5] introduced PGAN-
OPC, a mask optimization framework utilizing Conditional Generative Adversarial
Networks (CGANs). To address the inherent instability in GAN training, an ILT-
guided pre-training phase was incorporated, ensuring stable convergence and reducing
variability in image generation. Jiang et al. [8] proposed Neural-ILT, an end-to-end
framework built upon the U-Net architecture[12] with additional optimization lay-
ers, which enhances computational efficiency through CUDA acceleration and reduces
mask complexity. This model integrates an ILT correction layer and a complexity
refinement layer to optimize mask patterns. Similar to PGAN-OPC, Neural-ILT em-
ploys a pre-training phase to establish a robust foundation for target-mask mapping.
Compared to both PGAN-OPC and conventional ILT methods, Neural-ILT demon-
strates superior mask printability while significantly reducing mask complexity. Chen
et al. [7] introduced DevelSet, an improved level set-based ILT framework that com-
bines CUDA and DNN acceleration to enhance printability and enable rapid iterative
convergence. This approach refines the level set-based ILT algorithm by incorporating
a curvature term to minimize mask complexity while leveraging GPU capabilities to
overcome computational bottlenecks.
2
Building upon prior advancements in printability and computational efficiency,
this study seeks to further enhance performance by leveraging BNNs. We propose a
novel mask optimization framework, termed B-OPC, which integrates a U-Net-inspired
supervised learning approach with BNNs to achieve substantial reductions in compu-
tational cost. This innovation aims to streamline both model training and the mask
optimization process. While BNNs excel in reducing memory requirements and com-
putational overhead, a key challenge is preserving mask quality and minimizing the
information loss introduced by binarization. To address this, B-OPC employs a U-Net
architecture to ensure robust initialization and stable convergence. U-Net’s skip con-
nections are particularly critical, as they enable the preservation of low-level features,
such as edges, from earlier layers, bypassing the bottleneck and contributing directly
to the decoding process [12]. This mechanism ensures that fine-grained details are pre-
served, even within a binarized framework. Although advanced U-Net variants, such
as UNet++[14] and UNet3+[15], have been developed, their integration with BNNs
remains unexplored in the current literature. Given the sensitivity of BNNs training to
hyperparameter adjustments, the U-Net architecture offers a conservative yet effective
choice for achieving the desired balance between efficiency and performance.
2 Preliminaries
2.1 Lithography Simulation
The B-OPC framework generates optimized mask patterns, and to evaluate their ef-
fectiveness, it is essential to simulate the lithography process to produce corresponding
resist images. For this purpose, we utilize a lithography simulation model based on
Hopkins’ diffraction theory [16], which provides an approximation of the printed re-
sults. This model computes the aerial image I, which represents the light intensity
projected onto the wafer. The light intensity is derived by convolving the mask pattern
M with the optical kernels hk . The computation is expressed as follows:
2
N
X
I(x, y) = ωk |M(x, y) ⊗ hk (x, y)|2 (1)
k=1
where hk denotes the k-th optical kernel, and ωk represents its corresponding
weight. To balance computational efficiency and accuracy, we adopt an approxima-
tion of order Nk = 24, as suggested in[17]. This simplification enables the effective
evaluation of mask quality while maintaining manageable computational costs.
3
The L2 loss quantifies the fidelity of the nominal printed pattern to the target
design. It is defined as the squared Euclidean distance between the nominal printed
image Znom and the target image T:
where T represents the target image, and Znom denotes the printed image under
nominal process conditions.
The PVB measures the extent of variation in the printed pattern across different
process conditions, ±2% dose error in this case, reflecting the robustness of the de-
sign under real-world manufacturing variations. The PVB is defined as the squared
Euclidean distance between the printed images under maximum Zmax and minimum
Zmin process conditions:
where Zmax and Zmin represent the printed images under maximum and minimum
process conditions, respectively.
where xb represents the binarized variable and x the real-valued variable. The
binarization process enables efficient matrix multiplication through XNOR operations
and pop-counts, demonstrating a substantial reduction in computational demands.
However, the binarization process introduces challenges such as gradient mismatch
and information loss during training [18].
This approach enables gradient propagation and facilitates effective training. Dur-
ing backward propagation, real-valued gradients are accumulated for weight updates.
4
Figure 1: B-OPC model structure
To ensure compliance with the binarization constraint, weights are clipped between −1
and 1, and binarized values are recalculated. This STE-enabled method allows BNNs
to achieve significant efficiency while maintaining stable training dynamics. Addi-
tionally, binarization acts as a form of regularization, aiding in better generalization
[13].
3 Model: B-OPC
The objective of our DNN approach to OPC is to employ BNN to identify the opti-
mized mask Mop for a given layout target Zt , such that Mop closely approximates the
ground truth mask M .The dataset used in this study was sourced from the authors of
GAN-OPC[5]. It consist of 4,875 training instances synthesized based on design speci-
fications from existing 32nm M1 layout topologies, with each image sized at 256 × 256
pixels. Among these 4,875 design target and ground truth mask pairs, 4,400 instances
were utilized for training and validation, with a 9:1 split ratio, while 475 instances
were reserved for testing.
5
These skip connections concatenate features from the corresponding encoder layers,
preserving fine-grained details from the input image. The output layer consists of a
final quantized 2D transposed convolution with a single filter and hard tanh activation,
producing a 256 × 256 × 1 binary mask for OPC correction.
Binarization is implemented throughout the B-OPC model using the ste-sign quan-
tizer for both inputs and kernels in the convolutional layers, built using the Larq
framework[19]. This approach constrains weights and activations to binary values (-1
or +1), significantly reducing computational complexity and memory requirements.
To ensure the weights remain within the binary range during training, the weightclip
constraint is applied to the kernels. Kernel weights are initialized to ones, providing
a uniform starting point for the binarization process and enhancing training stability.
The hard tanh function serves as the activation throughout the network, approximat-
ing the binary step function while allowing gradient flow during backward propagation.
Together, these binarization techniques and activation functions enable the network
to effectively learn OPC corrections within the limitations of a binary representation,
achieving a trade-off between computational efficiency and performance.
4 Experimental Results
To evaluate the performance of the B-OPC model, we conducted experiments using
two distinct test cases. The first test case consists of 475 instances, separated from the
4,875 synthesized instances derived from the dataset provided by GAN-OPC [5]. Eval-
6
(a)
(b)
uating B-OPC on this test set provides an initial assessment of the model’s training
outcomes. The second test case comprises ten ICCAD 2013 contest test designs, which
are industrial M1 designs based on the 32nm technology node. This dataset serves as a
widely recognized benchmark in mask optimization research and has been extensively
used in previous studies [17, 5, 8, 7]. Evaluating B-OPC on this benchmark dataset
allows us to determine whether the model can effectively handle realistic circuit design
patterns. Furthermore, it enables a direct comparison with other DNN-based mask
optimization methods, providing insights into B-OPC’s competitiveness and practical
applicability.
7
Table 1: Comparison Lithography Simulation Result of Ground Truth
Masks and BOPC Generated Masks
(a)
(b)
(c)
(d)
Figure 3: (a)-(d) are four cases in test dataset. The images from
left to right represent the target image, ground truth mask, printed
image of ground truth mask, B-OPC mask, printed image of B-OPC
mask
An intriguing observation was made regarding the ground truth mask: certain de-
sign patterns are missing in the printed images derived from the ground truth masks,
but are accurately detected in the B-OPC masks. This phenomenon is evident in
the third and fourth rows of Figure 3. Specifically, small rectangular patterns, which
are separated and isolated from larger pattern regions in the design, are inadequately
compensated in the ground truth masks. These masks fail to include compensation
features around such isolated structures. In contrast, the masks generated by B-OPC
successfully address these limitations. B-OPC incorporates compensation features for
the isolated rectangular structures, ensuring they are preserved in the printed images.
This improvement demonstrates B-OPC’s capability to handle complex and isolated
design features more effectively than the ground truth masks, enhancing the overall
fidelity of the lithographic results.
8
Figure 4: Evaluation flow of ten ICCAD 2013 test design
The quantitative results after the ILT-refinement process are presented in Table 2.
The results demonstrate a significant improvement in L2 loss, with an average re-
duction to 60.1% of the initial value. However, the PVB exhibits a slight increase,
averaging 6.9% higher than the pre-refinement values. This increase in PVB can be
attributed to the successful printing of patterns that were previously omitted in the
coarse masks. Consequently, the total printed area becomes larger, leading to a higher
9
PVB measurement. Figure 5 provides visual examples of this phenomenon through
several cases from the benchmark dataset.
Interestingly, the lithography results from the first stage of B-OPC, as shown in
the third column of Figure 5, are comparable to the results produced by the U-Net
backbone in Neural-ILT, as depicted in Figure 4 of [8]. However, B-OPC offers signif-
icant advantages in terms of training simplicity and resource efficiency. For instance,
the U-Net structure in Neural-ILT requires pre-training for 20 epochs, which takes
approximately 19 hours on a single Titan V GPU [8]. In contrast, B-OPC completes
its training in 60 epochs within just 1 hour on an L4 GPU provided by Google Colab.
10
(a)
(b)
(c)
(d)
Figure 5: (a)-(d) are four cases in the benchmark. The image from
left to right represents the target image, the coarse mask, the printed
image of the coarse mask, the finetuned mask, the printed image of
the finetuned mask
11
5 Conclusion
This research proposes B-OPC, a Binary Neural Network(BNN)-based mask optimiza-
tion framework that demonstrates exceptional computational efficiency during model
training, requiring significantly less time and resources compared to conventional meth-
ods. By leveraging the inherent binary nature of mask patterns and design targets,
B-OPC achieves substantial reductions in memory usage and computational overhead
while maintaining high performance in semiconductor lithography. The experimen-
tal results demonstrate that B-OPC effectively compensates for diffraction effects on
specific design patterns. By incorporating a lightweight ILT-refinement process, B-
OPC achieves performance comparable to state-of-the-art methods such as Neural-
ILT, PGAN-OPC, and DevelSet, while requiring significantly less training time and
lower hardware resources.
References
[1] J. Kuang, W.-K. Chow, and E. F. Y. Young, “A robust approach for process
variation aware mask optimization,” in 2015 Design, Automation Test in Europe
Conference Exhibition (DATE), pp. 1591–1594, 2015.
[2] A. Awad, A. Takahashi, S. Tanaka, and C. Kodama, “A fast process variation
and pattern fidelity aware mask optimization algorithm,” in 2014 IEEE/ACM In-
ternational Conference on Computer-Aided Design (ICCAD), pp. 238–245, 2014.
[3] A. Poonawala and P. Milanfar, “Mask design for optical microlithography—an
inverse imaging problem,” IEEE Transactions on Image Processing, vol. 16, no. 3,
pp. 774–788, 2007.
[4] D. S. Abrams and L. Pang, “Fast inverse lithography technology,” in Optical
Microlithography XIX (D. G. Flagello, ed.), vol. 6154, p. 61541J, International
Society for Optics and Photonics, SPIE, 2006.
[5] H. Yang, S. Li, Y. Ma, B. Yu, and E. F. Y. Young, “Gan-opc: Mask op-
timization with lithography-guided generative adversarial nets,” in 2018 55th
ACM/ESDA/IEEE Design Automation Conference (DAC), pp. 1–6, 2018.
[6] G. Chen, W. Chen, Y. Ma, H. Yang, and B. Yu, “Damo: Deep agile mask op-
timization for full chip scale,” in 2020 IEEE/ACM International Conference On
Computer Aided Design (ICCAD), pp. 1–9, 2020.
[7] G. Chen, Z. Yu, H. Liu, Y. Ma, and B. Yu, “Develset: Deep neural level set for
instant mask optimization,” in 2021 IEEE/ACM International Conference On
Computer Aided Design (ICCAD), pp. 1–9, 2021.
[8] B. Jiang, L. Liu, Y. Ma, H. Zhang, B. Yu, and E. F. Young, “Neural-ilt: Migrat-
ing ilt to neural networks for mask printability and complexity co-optimization,”
12
in 2020 IEEE/ACM International Conference On Computer Aided Design (IC-
CAD), pp. 1–9, 2020.
[9] H. Yang and H. Ren, “Enabling scalable ai computational lithography with
physics-inspired models,” in 2023 28th Asia and South Pacific Design Automation
Conference (ASP-DAC), pp. 715–720, 2023.
[10] X. Liang, Y. Ouyang, H. Yang, B. Yu, and Y. Ma, “Rl-opc: Mask optimiza-
tion with deep reinforcement learning,” IEEE Transactions on Computer-Aided
Design of Integrated Circuits and Systems, vol. 43, no. 1, pp. 340–351, 2024.
[11] S. Zheng, H. Yang, B. Zhu, B. Yu, and M. Wong, “Lithobench: Benchmark-
ing ai computational lithography for semiconductor manufacturing,” in Advances
in Neural Information Processing Systems (A. Oh, T. Naumann, A. Globerson,
K. Saenko, M. Hardt, and S. Levine, eds.), vol. 36, pp. 30243–30254, Curran
Associates, Inc., 2023.
[12] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for
biomedical image segmentation,” in Medical Image Computing and Computer-
Assisted Intervention – MICCAI 2015 (N. Navab, J. Hornegger, W. M. Wells,
and A. F. Frangi, eds.), (Cham), pp. 234–241, Springer International Publishing,
2015.
[13] M. Courbariaux, I. Hubara, D. Soudry, R. El-Yaniv, and Y. Bengio, “Binarized
neural networks: Training deep neural networks with weights and activations
constrained to +1 or -1,” 2016.
[14] Z. Zhou, M. M. Rahman Siddiquee, N. Tajbakhsh, and J. Liang, “Unet++: A
nested u-net architecture for medical image segmentation,” in Deep Learning in
Medical Image Analysis and Multimodal Learning for Clinical Decision Support
(D. Stoyanov, Z. Taylor, G. Carneiro, T. Syeda-Mahmood, A. Martel, L. Maier-
Hein, J. M. R. Tavares, A. Bradley, J. P. Papa, V. Belagiannis, J. C. Nascimento,
Z. Lu, S. Conjeti, M. Moradi, H. Greenspan, and A. Madabhushi, eds.), (Cham),
pp. 3–11, Springer International Publishing, 2018.
[15] H. Huang, L. Lin, R. Tong, H. Hu, Q. Zhang, Y. Iwamoto, X. Han, Y.-W. Chen,
and J. Wu, “Unet 3+: A full-scale connected unet for medical image segmenta-
tion,” 2020.
[16] H. H. Hopkins, “The concept of partial coherence in optics,” Proceedings of the
Royal Society of London A: Mathematical, Physical and Engineering Sciences,
vol. 208, no. 1093, pp. 263–277, 1951.
[17] J.-R. Gao, X. Xu, B. Yu, and D. Z. Pan, “Mosaic: Mask optimizing solution
with process window aware inverse correction,” in 2014 51st ACM/EDAC/IEEE
Design Automation Conference (DAC), pp. 1–6, 2014.
[18] H. Qin, R. Gong, X. Liu, X. Bai, J. Song, and N. Sebe, “Binary neural networks:
A survey,” CoRR, vol. abs/2004.03333, 2020.
[19] L. Geiger and P. Team, “Larq: An open-source library for training binarized
neural networks,” Journal of Open Source Software, vol. 5, no. 45, p. 1746, 2020.
[20] J. Jiang, “Lithobench: Benchmarking ai computational lithography for semi-
conductor manufacturing.” https://github.com/shelljane/lithobench, 2023. Ac-
cessed: 2024/6/7.
13