Book a demo

Cut patent&paper research from weeks to hours with PatSnap Eureka AI!

Try now

Machine learning fixes TCAD simulation convergence issues

Machine Learning for TCAD Simulation Convergence — PatSnap Insights
Semiconductor Innovation

Machine learning is reshaping how engineers tackle the most stubborn bottleneck in semiconductor device development: TCAD simulation convergence. From graph neural networks that seed Newton-Raphson solvers with physically consistent starting points, to active learning frameworks that cut TCAD query counts to a minimum, the hybrid ML-TCAD paradigm is delivering speedups that were unthinkable a decade ago.

PatSnap Insights Team Innovation Intelligence Analysts 11 min read
Share
Reviewed by the PatSnap Insights editorial team ·

Why TCAD Convergence Fails at Advanced Nodes

Traditional TCAD simulation fails to converge at advanced technology nodes primarily because iterative numerical solvers—particularly Newton-Raphson loops coupling Poisson’s equation with carrier transport equations—accumulate errors that compound as device geometries shrink and three-dimensional complexity grows. As documented by Samsung Electronics (2021), the core problem is twofold: prohibitive computational cost and recurrent convergence errors in three-dimensional TCAD simulation. The same study explicitly states that complete replacement of TCAD with deep learning has not yet been achieved, framing ML instead as a convergence stabilizer within a restructured hybrid system.

122,000×
Max simulation speedup vs. traditional TCAD (Korea University, 2022)
R²=0.99
Neural surrogate accuracy on FinFET structures
10⁷×
Faster MOS calibration vs. direct TCAD (Xiangtan University, 2024)
100%
Transient convergence with ISRU activation in ML compact models
20+
Patents and publications analysed, spanning 2014–2026

The mechanics of convergence failure are well understood: as the Newton-Raphson solver propagates corrections across coupled differential equations, small numerical inconsistencies in the initial solution guess can cause oscillation or divergence rather than convergence to a self-consistent solution. A patent from Peking University Shenzhen Research Institute (2014) addresses this directly by recording M historical input-output iteration pairs and using weighted linear combinations to predict the next iteration input—one of the earliest verified instances of data-driven convergence acceleration applied to semiconductor device physics solvers.

The problem intensifies with device scaling. As FinFETs give way to gate-all-around (GAA) structures and 3D NAND flash memory stacks grow taller, the simulation mesh required to resolve critical dimensions expands dramatically, multiplying both the computational cost per iteration and the probability that any given iteration will fail to converge. According to research from IEEE, the computational demands of full-physics 3D device simulation have grown faster than available compute resources at each successive technology node, making algorithmic acceleration—rather than hardware scaling alone—a practical necessity.

Samsung Electronics (2021) reported that complete replacement of TCAD with deep learning has not yet been achieved, and instead proposed a hybrid restructuring algorithm that enables deep learning and TCAD to complement each other while fully resolving convergence errors in three-dimensional TCAD simulation.

What is TCAD?

Technology Computer-Aided Design (TCAD) refers to software tools that simulate the physics of semiconductor device fabrication and operation using numerical methods. TCAD solvers model carrier transport, electrostatics, thermal effects, and quantum phenomena to predict device characteristics without requiring physical fabrication—making convergence of these iterative solvers a critical bottleneck in semiconductor R&D.

The dataset analysed for this article encompasses more than 20 patents and peer-reviewed publications spanning 2014 to 2026, drawn from institutions including Samsung Electronics, Synopsys, Gwangju Institute of Science and Technology (GIST), Tsinghua University, Seoul National University, Korea University, and Lam Research. A unifying observation across the entire dataset is that pure TCAD replacement by ML remains elusive—hybrid ML-TCAD complementarity is the prevailing architectural philosophy.

Surrogate Models and Active Learning: Doing More With Fewer Simulations

The central challenge in ML-based TCAD acceleration is data acquisition: sufficient TCAD simulation results must be generated to train predictive models, yet each simulation is expensive and may itself fail to converge. Active learning frameworks solve this by intelligently selecting the most informative simulation points, eliminating redundant runs that would be required by exhaustive sampling.

Alsemi (2024, 2025) has filed a cluster of patents disclosing an active learning agent that queries TCAD simulations only for voltage-current or voltage-capacitance pairs that maximize model informativeness, then trains a neural network compact model on the returned results. A related filing discloses a reinforcement learning agent for optimizing process variables using the neural compact model, establishing a closed-loop pipeline from TCAD data to compact model to process optimization without manual intervention.

“Speedups exceeding 122,000× compared to traditional simulators have been validated for FinFET structures at R² = 0.99 accuracy—establishing a benchmark for practical deployment viability.”

A complementary Bayesian approach is documented by the Taiwan Semiconductor Research Institute (2022) in the context of semiconductor laser annealing. The proposed meta-learner adjusts the hybridization ratio between TCAD and ML when selecting the next sampling point, resulting in significantly lower mean square error during the first 100 sampling steps compared to pure ML approaches. This is particularly relevant for processes where experimental iteration is prohibitively expensive, as reported in research indexed by Nature on advanced semiconductor fabrication process windows.

Figure 1 — ML-TCAD Simulation Speedup Benchmarks Across Methods and Institutions
Machine Learning TCAD Simulation Speedup Comparison: Neural Surrogates, Random Forest, CNN-NEGF 10³× 10⁵× 10⁷× Speedup (log scale) 122,000× FinFET Neural Surrogate (Korea Univ.) 10⁷× MOS RF Model Calibration (Xiangtan Univ.) ~100× CNN-NEGF Acceleration (Kobe Univ.) ~10× Historical Info Iteration (PKU, 2014) * Speedup values on logarithmic scale. CNN-NEGF and historical iteration values are representative estimates from qualitative descriptions in source papers.
Neural surrogate models for FinFET simulation (Korea University, 2022) deliver speedups exceeding 122,000× at R² = 0.99 accuracy. MOS transistor calibration using random forest surrogates (Xiangtan University, 2024) reports approximately 10⁷× faster computation than direct TCAD. All values plotted on a logarithmic scale.

For process simulation model calibration, Lam Research has filed multiple patent generations on optimizing process simulation model parameters by minimizing cost functions derived from CD-SEM, optical scatterometry, and transmission electron microscopy measurements. The optimization loop adjusts pre-process profiles to match post-process metrology, representing a data-driven calibration approach that reduces manual tuning iterations across advanced patterning processes.

Tsinghua University has extended active learning into transistor design optimization (2025) using Gaussian process regression as a surrogate model and differential evolution to generate candidate process parameter sets. Only candidates with high predicted performance and high uncertainty are forwarded to TCAD for simulation, with results iteratively fed back to update the surrogate. This approach directly targets the Pareto front of multi-objective transistor performance optimization, substantially shortening the development cycle.

Taiwan Semiconductor Research Institute (2022) demonstrated that a meta-learned Bayesian surrogate that adjusts the hybridization ratio between TCAD and ML when selecting sampling points achieves significantly lower mean square error during the first 100 sampling steps compared to pure ML approaches for semiconductor laser annealing process optimization.

Explore the full patent landscape for ML-TCAD convergence acceleration with PatSnap Eureka.

Search ML-TCAD Patents in PatSnap Eureka →

Neural Architecture Choices and Their Convergence Impact

The choice of neural network architecture directly determines both the accuracy of ML-TCAD hybrid systems and the nature of the convergence improvement they deliver. Four distinct architectural paradigms appear across the dataset, each targeting a different failure mode in the TCAD simulation pipeline.

Graph Neural Networks for Initial Solution Generation

Graph neural networks (GNNs) address convergence at its root cause: a poor initial solution guess. Gwangju Institute of Science and Technology (GIST, 2023) patented a system that derives an area graph from the device structure file, processes it through a graph neural network to classify the device type, and generates a compact-model-derived initial solution. This initial solution is passed to the TCAD simulator before the Newton-Raphson loop begins, dramatically reducing the number of iterations required to reach convergence by starting from a physically plausible estimate. A hierarchical variant (GIST, 2021) employs a higher-level DNN that provides an initial solution hint to a lower-level device simulator, enabling multi-level convergence guidance.

Convolutional Neural Networks for Quantum Transport Acceleration

The nonequilibrium Green’s function (NEGF) method is among the most computationally intensive TCAD kernels, required for accurate quantum transport simulation in nanoscale FETs. Kobe University (2020) demonstrated that a convolutional autoencoder trained to predict carrier density and local quantum capacitance distributions from input potential distributions can be substituted into the NEGF self-consistent loop alongside Poisson’s equation, yielding accurate potentials across a range of gate lengths in significantly shorter computation time than conventional NEGF. This substitution is particularly significant because NEGF convergence failures are common and expensive to diagnose.

Figure 2 — ML Architecture Selection by TCAD Convergence Failure Mode
Neural Network Architecture Selection for TCAD Simulation Convergence: GNN, CNN, LSTM, RNN Roles GNN Initial solution generation ↓ Newton-Raphson iters CNN NEGF quantum transport kernel ↓ NEGF compute cost LSTM Time-step control in PTA circuits ↓ DC divergence failures RNN Sequential process step emulation ↓ Per-step TCAD calls GIST 2023 Kobe Univ. 2020 Huada Empyrean 2023 Samsung 2021 Each architecture targets a distinct convergence failure mode in the TCAD pipeline
Four neural architectures address distinct TCAD convergence failure modes: GNNs reduce Newton-Raphson iterations via better initial solutions; CNNs substitute expensive NEGF kernels; LSTMs prevent pseudo-transient analysis divergence; RNNs bypass per-step TCAD calls in sequential process simulation.

LSTM Networks for Time-Step Control in Circuit Simulation

One of the most common convergence failure modes in TCAD-linked circuit simulation is pseudo-transient analysis (PTA) divergence, caused by time steps that are either too large (causing numerical instability) or too small (causing excessive conservatism and runtime). Huada Empyrean Software (2023) proposed an LSTM-based method exploiting the network’s temporal memory to predict future optimal time steps from a coarse-and-fine-grained hybrid sampling strategy. By learning from the history of solver states, the LSTM avoids both failure modes—directly targeting DC convergence reliability in production SPICE and TCAD-linked simulation environments.

Recurrent Neural Networks for Sequential Process Emulation

Samsung Electronics (2021) disclosed a simulation system built around a recurrent neural network consisting of process emulation cells arranged in time series. Each cell receives the previous output profile, target profile, and process condition information, and generates the current output profile using time-series causal relationships encoded as prior knowledge. This architecture mirrors the sequential physics of multi-step semiconductor manufacturing, enabling step-by-step profile prediction that bypasses TCAD at each individual process step—a particularly powerful approach for processes with many sequential operations, such as multi-patterning lithography stacks.

Gwangju Institute of Science and Technology (GIST, 2023) patented a graph neural network system that derives an area graph from a semiconductor device structure file, classifies device type, and generates a compact-model-derived initial solution to pass to the TCAD simulator—reducing Newton-Raphson iterations by starting from a physically plausible initial condition.

Optimizing Simulator Settings and Compact Model Convergence

Beyond replacing TCAD kernels, machine learning has been applied to optimizing the configuration of TCAD simulators themselves—mesh density, solver tolerance, stepping strategy—tuning parameters that critically affect both convergence and accuracy. Synopsys (2022) filed the most direct patent expression of this concept, disclosing an ML framework that predicts optimized TCAD simulator system settings for future simulation executions by learning from past simulation runs which settings led to convergence and which did not.

Samsung Electronics (2024) addressed the specific problem of regression model prediction failure in neural network-based semiconductor characteristic prediction. When consistency of the trained model falls below a target threshold, the system identifies the consistency reduction factors—regions of the input space where the model fails—and retrains on additional sample data targeted at those regions. This adaptive retraining loop is conceptually equivalent to active learning but applied to model consistency rather than output uncertainty, and represents a production-grade reliability mechanism for deployed ML-TCAD systems.

Key finding: Activation function selection is critical for SPICE convergence

National Yang-Ming Chiao-Tung University (2021) evaluated sigmoid, tanh, ReLU, and inverse square root unit (ISRU) activation functions in ML compact models and found that ISRU activation achieves 100% convergence in transient analysis—substantially outperforming all other activations. The presence of gate-to-source and gate-to-drain capacitances in the model further benefits convergence.

The impact of ML compact model design on SPICE convergence is rigorously analysed in a 2021 study from National Yang-Ming Chiao-Tung University, Taiwan. The study evaluates multiple activation functions and finds that ISRU activation achieves 100% convergence in transient analysis, substantially outperforming sigmoid, tanh, and ReLU. This result has direct implications for how ML compact models should be architected for integration into SPICE engines—a finding that has been cited by subsequent compact model research tracked by OECD technology foresight reports on semiconductor design automation.

A MOS transistor simulation model calibration patent from Xiangtan University (2024) combines a classifier, regressor, and surrogate model to calibrate TCAD parameters, reporting a computation speed approximately 10⁷ times faster than direct TCAD simulation. The method trains a random forest model to replace expensive TCAD calls during the iterative calibration loop—demonstrating that ensemble methods, not just deep neural networks, have a role in the ML-TCAD acceleration stack. This aligns with standards for model validation published by ISO for simulation-based design processes.

National Yang-Ming Chiao-Tung University (2021) found that inverse square root unit (ISRU) activation functions in machine learning compact models achieve 100% convergence in transient SPICE analysis, outperforming sigmoid, tanh, and ReLU activations. The presence of gate-to-source and gate-to-drain capacitances in the model further benefits convergence.

Analyse compact model convergence patents and identify white spaces with PatSnap Eureka’s AI-powered search.

Analyse Compact Model Patents in PatSnap Eureka →

Who Is Building the ML-TCAD Stack: Key Players and Trends

Samsung Electronics and Synopsys are the most prolific patent holders in ML-TCAD convergence research, while academic contributors from Korea, Taiwan, and China dominate the peer-reviewed literature. The dataset spanning 2014 to 2026 reveals a clear progression from offline-trained neural network surrogates toward closed-loop multi-agent systems.

Samsung Electronics is the dominant patent filer, with multiple patents on semiconductor design automation systems integrating ML with TCAD simulation data. Samsung also contributes foundational literature on 3D TCAD restructuring and neural network-based semiconductor characteristic prediction, including a 2024 patent that trains a deep learning model on TCAD simulation result data to generate target prediction data with adaptive retraining when model consistency drops below threshold.

Synopsys contributes both the ML-optimized simulator settings patent (2022) and a machine-learning-driven prediction framework for integrated circuit design (EP, 2024), establishing it as the leading EDA vendor in this space. Synopsys also published foundational work in 2019 on machine learning-assisted modeling frameworks for design-technology co-optimization (DTCO), predating many of the patent filings now citing that work.

Alsemi (주식회사 알세미) has filed a focused cluster of active learning patents for neural compact model generation, representing a Korean startup effort targeting the TCAD-to-compact-model pipeline with reinforcement learning agents for process variable optimization.

Gwangju Institute of Science and Technology (GIST) holds active patents on graph neural network-based device simulation initialization, covering both the system architecture and hierarchical DNN approaches for initial solution generation.

Lam Research focuses on process simulation model calibration using metrology-driven cost functions, filing multiple patent generations on CD-SEM-based calibration methodology across advanced patterning processes.

Figure 3 — ML-TCAD Innovation Timeline: From Offline Surrogates to Multi-Agent Systems (2014–2026)
ML-TCAD Innovation Timeline 2014–2026: From Data-Driven Iteration to Multi-Agent Reinforcement Learning 14 2014 Historical info iteration method (PKU Shenzhen) 19 2019–2021 Offline neural surrogates & DTCO frameworks 22 2022–2023 Active learning & adaptive retraining loops 24 2024–2026 Multi-agent RL & LLM integration for param. optim.
The ML-TCAD innovation trajectory spans from simple data-driven iteration methods (2014) through offline neural surrogates (2019–2021) and active learning loops (2022–2023), toward multi-agent reinforcement learning and large language model integration for parameter optimization (2024–2026).

From a trend perspective, the dataset shows clear progression: from simple neural network surrogates trained offline (2019–2021), through active learning and adaptive retraining loops (2022–2023), toward multi-agent reinforcement learning and large language model integration for parameter optimization (2024–2025). This rapid expansion of the ML-TCAD solution space is tracked by innovation intelligence platforms and reported in semiconductor technology roadmaps published by bodies such as WIPO in its annual global innovation index for semiconductor technologies.

The ML-TCAD research dataset spanning 2014 to 2026 shows a progression from simple neural network surrogates trained offline (2019–2021), through active learning and adaptive retraining loops (2022–2023), toward multi-agent reinforcement learning and large language model integration for parameter optimization (2024–2025), with Samsung Electronics and Synopsys as the most prolific patent holders.

Frequently asked questions

Machine learning for TCAD simulation convergence — key questions answered

Still have questions? Let PatSnap Eureka answer them for you.

Ask PatSnap Eureka for a Deeper Answer →

References

  1. Restructuring TCAD System: Teaching Traditional TCAD New Tricks — Samsung Electronics Data and Information Technology Center, 2021
  2. Method and computing device for improving the generation efficiency of neural compact models — Alsemi Co., Ltd., 2025
  3. Active learning methods and computing devices for neural compact models — Alsemi Co., Ltd., 2024
  4. Machine learning method and framework for optimizing setups for accurate, speedy and robust TCAD simulations — Synopsys, Inc., 2022
  5. Circuit convergence study using machine learning compact models — National Yang-Ming Chiao-Tung University, Taiwan, 2021
  6. Meta-Learned and TCAD-Assisted Sampling in Semiconductor Laser Annealing — Taiwan Semiconductor Research Institute, 2022
  7. Accelerating DC Circuit Simulation through Feature Selection and LSTM-Based Time-Step Control — Huada Empyrean Software Co., Ltd., 2023
  8. Simulator acceleration and inverse design of fin field-effect transistors using machine learning — Korea University, 2022
  9. Semiconductor device simulation system and semiconductor device simulation method — Gwangju Institute of Science and Technology, 2023
  10. Electronic device, method, and computer readable medium for simulation of semiconductor device — Gwangju Institute of Science and Technology, 2021
  11. Acceleration of nonequilibrium Green’s function simulation for nanoscale FETs by applying convolutional neural network model — Kobe University, 2020
  12. Modeling method of neural network for simulation in semiconductor design process — Samsung Electronics Co., Ltd., 2024
  13. Method of predicting characteristic of semiconductor device and computing device performing the same — Samsung Electronics Co., Ltd., 2023
  14. Semiconductor design automation system and computing system including the same — Samsung Electronics Co., Ltd., 2023
  15. Method for efficiently optimizing process variables in semiconductor devices using neural compact model — Alsemi Co., Ltd., 2024
  16. Process simulation model calibration using CD-SEM — Lam Research Corporation, 2025
  17. Machine-learning driven prediction in integrated circuit design — Synopsys, Inc., 2024
  18. Simulation system for semiconductor process and simulation method thereof — Samsung Electronics Co., Ltd., 2021
  19. Active learning strategy-based transistor key characteristic optimization method and system — Tsinghua University, 2025
  20. A machine learning-based MOS transistor simulation model calibration method — Xiangtan University, 2024
  21. A rapid simulation method for semiconductor nanodevices based on historical information — Peking University Shenzhen Research Institute, 2014
  22. New-Generation Design-Technology Co-Optimization (DTCO): Machine-Learning Assisted Modeling Framework — Synopsys, Inc., 2019
  23. IEEE — Institute of Electrical and Electronics Engineers (authority source)
  24. WIPO — World Intellectual Property Organization (authority source)
  25. PatSnap Insights — Innovation Intelligence Blog

All data and statistics in this article are sourced from the references above and from PatSnap‘s proprietary innovation intelligence platform.

Your Agentic AI Partner
for Smarter Innovation

PatSnap fuses the world’s largest proprietary innovation dataset with cutting-edge AI to
supercharge R&D, IP strategy, materials science, and drug discovery.

Book a demo