Abstract
Optical emission spectroscopy (OES) of helium (He) line intensities has been used to measure the electron density, ne, and temperature, Te, in various plasma devices. In this study, a neural network with five hidden layers is introduced to model the relation between the OES data and ne/Te from laser Thomson scattering in the linear plasma device Magnum-PSI and compared to multiple regression analysis. It is shown that the neural network reduces the residual errors of prediction values (ne and Te) less than half those of the multiple regression analysis in the ranges of 2 × 1018<ne<8×1020 m−3 and 0.1<Te<4eV. We checked two different data splitting methods for training and validation data, i.e., with and without considering the unit of discharge. A comparison of the splitting methods suggests that the residual error will decrease to ∼10% even for a new discharge data when accumulating a sufficient data set.
Original language | English |
---|---|
Article number | 101281 |
Journal | Nuclear Materials and Energy |
Volume | 33 |
DOIs | |
State | Published - Oct 2022 |
Funding
This work was supported in part by a Grant-in-Aid for Scientific Research 19H01874 and Fund for the Promotion of Joint International Research 17KK0132 and 21KK0048 from the Japan Society for the Promotion of Science (JSPS) , and NIFS collaboration research program ( NIFS22KIIH006 ).
Keywords
- Helium
- Machine learning
- Neural network
- Optical emission spectroscopy
- Plasma