Abstract
The helium line intensity ratio (LIR) with the help of a collisional radiative (CR) model has long been used to measure the electron density, n e , and temperature, T e , and its potential and limitations for fusion applications have been discussed. However, it has been reported that the CR model approach leads to deviations in helium-hydrogen mixed plasmas and/or recombining plasmas. In this study, a machine learning (ML) aided LIR method is used to measure n e and T e from spectroscopic data of helium-hydrogen mixed recombining plasmas in the divertor simulator Magnum-PSI. To analyze mixed plasmas, which have more complex spectral shapes, the spectroscopy data were used directly for training instead of separating the intensities of each line. It is shown that the ML approach can provide a robust and simpler analysis method to deduce n e and T e from the visible emissions in helium-hydrogen mixed plasmas.
Original language | English |
---|---|
Article number | 105005 |
Journal | Plasma Physics and Controlled Fusion |
Volume | 66 |
Issue number | 10 |
DOIs | |
State | Published - Oct 2024 |
Funding
This work was supported in part by a Fund for the Promotion of Joint International Research 17KK0132 and 21KK0048, and NIFS collaboration research program (NIFS22KIPP002, NIFS23HDAF011, NIFS23KIPH026).
Funders | Funder number |
---|---|
National Institute for Fusion Science | NIFS23KIPH026, NIFS23HDAF011, NIFS22KIPP002 |
Keywords
- divertor simulator
- helium
- machine learning
- neural network
- spectroscopy