Tian‑Shuai Shang · Jian Li · Zhong‑Ming Niu
Abstract Nuclear charge density distribution plays an important role in both nuclear and atomic physics, for which the two-parameter Fermi (2pF) model has been widely applied as one of the most frequently used models. Currently, the feedforward neural network has been employed to study the available 2pF model parameters for 86 nuclei, and the accuracy and precision of the parameter-learning effect are improved by introducing A1∕3 into the input parameter of the neural network. Furthermore, the average result of multiple predictions is more reliable than the best result of a single prediction and there is no significant difference between the average result of the density and parameter values for the average charge density distribution. In addition, the 2pF parameters of 284 (near) stable nuclei are predicted in this study, which provides a reference for the experiment.
Keywords Charge density distribution · Two-parameter Fermi model · Feedforward neural network approach
Starting from when the nucleus was discovered by Rutherford [1], its charge density distribution has been studied because it is critical for analyzing the nuclear structure. The nuclear charge density distribution provides direct information regarding the Coulomb energy of the nucleus, which allows it to be used to calculate the charge radius. Additionally, studies regarding high-momentum tails (HMT) caused by short-range correlations (SRC) found that the percentage of tails is closely related to the distribution of the nuclear charge density [2, 3]. An accurate estimation of the neutron and proton density distributions is crucial for the study of asymmetric nuclear matter in nuclear Astrophysics. The nuclear symmetry energy and its density-dependence play an important role in understanding the physics of several terrestrial nuclear experiments and astrophysical observations [4, 5]. Considering configurational information entropy(CIE), the charge distributions in projectile fragmentation reactions may be good probes for determining the thickness of the neutron skin of neutron-rich nuclei [6], among other characteristics, such as the density dependence of the symmetry energy [7, 8]. In contrast, the charge density distribution also has a high status in atomic physics [9, 10]. For example, if the charge density distribution in the nucleus is known, the deformation of the nucleus can be calculated and its influence on the electrons in the atom can then be determined [11, 12].
Hofstadter measured the charge density of protons in the 1950s and described the density distribution of certain nuclei based on the findings [13]. Thus far, electron scattering experiments (elastic and inelastic) have become an effective method for measuring the nuclear structure [14—17].After obtaining the density distribution data, the following two methods can be used to describe the shape of the charge density: a model-dependent analysis (e.g., two/three parametric Fermi models and two/three parametric Gaussian models), and model-independent analysis (e.g., Fourier Bessel and sum-of-Gaussian analyses) [18]. However, the experimental data regarding nuclear charge density remains limited. Considering the model-related analysis method as an example, fewer than 300 nuclei with charge density parameters were confirmed [19—21], which are mainly concentrated near stable nuclei.
In recent decades, several microscopic models of nuclear structure have been successfully established, nearly all of which can calculate the density distribution information,such as the ab-initio (Green’s function Monte Carlo method[22], self-consistent Green’s function method [23], coupled-cluster method [24], lattice chiral effective field theory[25], and Nocore shell model) and the density functional theory (DFT) [26]. Both of these can accurately describe the nuclear ground-state properties. However, as the nuclear mass number increases, the expansion of the configuration space limits the calculation range of ab-initio and the shell models. The systematicity of the calculation and accurately describing is a significant challenge. The density functional theory, such as the Skyrme-Hartree-Fock (SHF) method[27—29] and Covariant Density Functional Theory (CDFT)[30—36], are the most widely used and effective microscopic models for studying nuclear properties. The DFT,with a small number of parameters, allows a very successful description of the ground-state properties of nuclei all over the nuclear chart [26]. Although its calculation range on the nuclear chart is far beyond that of ab-initio and shell models, its prediction of the charge density distribution is occasionally inaccurate.
Compared to the aforementioned microscopic theoretical models, empirical models are more commonly used to describe the distribution of the nuclear charge density, such as the Fermi and Gaussian models in the model-dependent analysis method, among which the two-parameter Fermi(2pF) model is one of the most widely used. 2pF can describe the stability of the central density of larger nuclei and demonstrates the exponential decay of the surface density. More importantly, the 2pF model only requires two parameters to describe the nuclear charge density and is easy to use. However, only a few nuclei have the 2pF parameters according to Refs. [19—21], and other nuclei need to be extrapolated from existing experimental data. This can be easily performed using machine learning methods.
Machine Learning (ML) utilizes computers to simulate or achieve human learning activities. It is one of the most intelligent and cutting-edge research fields in artificial intelligence (AI). In recent decades, the prodigious development of machine learning applications has impacted several fields,such as image recognition [37, 38] and language translation[39, 40]. Many algorithms have been developed for machine learning to closely resemble the human mind, the core of which is the Back Propagation (BP) algorithm, which is the most powerful and popular machine learning tool. Other algorithms, such as the Decision Tree (DT), Naive Bayesian Model (NBM), SVM and Cluster have also been used in several areas, providing powerful tools for particle physics [41—43] and condensed matter physics [44, 45], among others.
In the 1990s, ML along with neural networks was applied to the modeling of observational data in nuclear physics [46,47] and have been widely used in various fields, such as to determine the ground state properties of nuclei including the nuclear mass (binding energy), stability, separation energy, and branching ratio of radioactive decay, among others [46—56]. Others include the excited state [57—60],charge radius [61—63],αdecay [64, 65],βdecay [66—68],magnetic moment [69], nuclear reactions and cross sections[70—74], nuclear structural data [75—79], giant dipole resonance [80],βdecay one-neutron emission probabilities [81],density functionals for nuclear systems [82], and nuclear data evaluation [83]. Among these, the feedforward neural network (FNN) is the most widely used. [46, 47, 50—54, 59,60, 63, 68]. This neural network demonstrates a significant learning ability because it can learn any function by adjusting the appropriate hyperparameters. Therefore, the FNN is adopted to study the 2pF experimental data and for predictions of other nuclei.
The basic formulas of the FNN approach are presented in Sect. 2, the prediction results of the charge density distribution are discussed in Sect. 3, and the summary and perspectives are presented in Sect. 4.
In most cases, a two-parameter Fermi distribution:
is assumed for the charge distribution. The parametercis the half-density radius andzis the diffuseness of the nuclear surface.N0is the normalization factor that satisfies the following:
whereZis the number of protons.
The feedback neural network can be categorized under the ML sub-area. FNN mimics human brain functionality to provide outputs as a consequence of the input computation. It is composed of processing units called neurons, which have an adaptive synaptic weights [60]. The FNN framework is illustrated in Fig. 1, which is a multilayer neural network consisting of an input layer, hidden layers, and an output layer. The number of hidden layers can vary and the neurons are fully connected between the layers.
The output of each layer of the neural network is denoted as [a1,a2,a3,…,an] , wherea1is the input of the network andanis the output of the network, and the number of each layer’s neurons are labeled as [N1,N2,N3,…,Nn] . For the hidden layer, the outputaiis calculated using the following formula:
whereWiis the weight matrix between thei-1th andith layers with shapes ofNi×Ni−1, andbiis the bias vector of theith layer. The activation functionfperforms a nonlinear mapping of the input, which is an important cause of FNN fitting most functions. In this study, the activation functions of hidden layers are considered the hyperbolic tangent, tanh:
Fig. 1 (Color online) A schematic diagram of a neural network with four input variables, two output variables, and two hidden layers (five neurons H =5 in each hidden layer). There are also four input and two output variables
which is used to quantify the difference between the model predictionsypreand the experimental valuesytar. Here,Nsdenotes the size of the training set. The learning process minimizes the loss function via a proper optimization method.A back-propagation algorithm by Levenberg-Marquardt[84, 85] was used to train the FNN. The FNN modifies its weights until an acceptable error level between the predicted and the desired outputs is achieved. The stochastic gradient descent (SGD) method [86] is used in this study to obtain the optimal parametersWiandbiin the network. The SGD is a popular alternative to the gradient descent (GD) method,which is one of the most widely used training algorithms.
Because the charge radius of the nucleus can be described byR=r0Z1∕3in most cases [87], consideringZ1∕3in the input will help fit the parametercof 2pF. The following alternative formula for the radius has been occasionally used in other previous studies[13, 88, 89]:R=r0A1∕3,whereAis the mass number. Therefore,A1∕3has inputs apart fromZ,N,Z1∕3to ascertain the effect of this additional input. For simplicity, FNN-I3 and FNN-I4 are used to represent the FNN approaches withx=(Z,N,Z1∕3) andx=(Z,N,Z1∕3,A1∕3) , respectively. To simultaneously predict the two parameterscandzin 2pF, FNN with a double hidden layer structure is adopted.
The experimental data for 2pF are obtained from Refs.[19—21]. There are 86 nuclei remaining and their experimental 2pF data are recorded in the dataset. If the 2pF parameter of a nucleus is obtained from multiple independent experiments, the most recent data will be used.
To obtain reliable results, the neural network is repeatedly trained (1000 times), and each training used randomly divided training and validation sets, among which the training set accounted for 80% of the dataset (86 nuclei) and the validation set accounted for 20% . Moreover, at the beginning of each training session, the network parameters are randomly reinitialized. After each training, FNN-I3 and FNNI4 obtain the prediction results of parameterscandzon the validation set, which can be compared with the experimental results from Refs. [19—21] to obtain a mean-squared-error(MSE).
Table 1 lists the MSE statistics of the parametercandzpredictions of the FNN-I3 and FNN-I4 approaches for the training and validation sets. As the result of FNN is affected by parameter initialization, certain training results deviate far from the experimental value. In this study, the following upper limits of MSE are set for the indicated parameters:0.2 for parameterc, and 0.005 for parameterz. Note, the two nuclei (98In and102Sb ) that are used in the dataset are distant from the others on the nuclide chart and are unstable.Removing these two nuclei from the dataset will not affect the results by more than 0.002 fm2and the relative error is less than 5% . These errors can be considered the result of removing the data from the dataset and do not represent any particularities of the two nuclei. Results with an MSE
Table 1 MSE statistics of parameters c and z on the training and validation sets; the datasets are randomly divided. SD represents the standard deviation calculated using Eq. (6)
The count indicates the number of results whose error is within the acceptable range (0.2 forcand 0.005 forz)greater than the upper limit are eliminated. The number of accepted predictions is shown in the last column. The standard deviation (SD) is also shown in the table, which can be obtained as follows:
whereNdenotes the number of predictions adopted.ypre
iindicates the predicted values andytaris the target value.
For parameterc, the mean squared deviation of FNN-I3 is clearly significantly larger than that of FNN-I4, and the SD of FNN-I4 is smaller. Compared to FNN-I3, the mean MSE of FNN-I4 is reduced from 0.03913 to 0.02408 fm2on the training sets and from 0.03655 to 0.02286 fm2on the validation sets, which is significantly precise for the predictions of parameterc. The nuclear charge radius is closely related to the mass and proton numbers. Therefore, the aforementioned results indicate that the FNN approach is reliable to improve the accuracy of the nuclear charge density distribution predictions based on experimental data by including the known effects of physics in the input layer. Note, different combinations of the physical quantities are used as the inputs,and it is found that parametercsignificantly correlated withA1∕3andZ1∕3(Pearson correlation coefficients ofA1∕3−candZ1∕3−care 0.9953 and 0.9969, respectively), whereas parameterzis not sensitive to it (Pearson correlation coefficients are — 0.1625 forA1∕3−cand — 0.1609 forZ1∕3−c).Therefore, the result of parameterzis not considered in the following analysis.
To further evaluate the prediction ability of the FNN method, the MSE statistics of parametercfor the fixed training and validation sets are demonstrated in Table 2. Thecolumns in Table 2 indicate the same meaning as those in Table 1. As the value of parametercranges from 2.4 to 7.0 fm, the FNN approach can be considered to be reliable.
Table 2 Table caption Mean squared error statistics of parameter c on the training and validation sets; the dataset is fixed
Figure 2 presents the prediction of the charge density distribution of certain nuclei with a fixed training and validation set. The average charge density distributions are obtained by applying the 100 FNN models with the best performance on the training set, which indicates the smallest loss function value, and applied it to the validation set to obtain the average charge density distributions. The error bands are obtained from the standard deviation of the density distribution values of the 100 FNN models. Figure 2 clearly demonstrates that the error bands of FNN-I4 are significantly narrower than those of FNN-I3, indicating that FNN-I4 has a higher precision than FNN-I3. Moreover, in most cases,the average prediction distributions of FNN-I4 are closer to the density distribution obtained using the experimental parameters compared to those of FNN-I3. In conclusion,FNN-I4 has a higher accuracy than FNN-I3.
To illustrate the rationality of using the average density distribution as the prediction result, we compare the average result of multiple predictions and the result with the minimum (maximum) error on the validation set, as shown in Fig. 3. The 2pF model uses parameters to control the charge density distribution; in addition to averaging the charge density distribution curve to obtain the average predicted density distribution, it can be obtained by averaging the predicted parameters. Thus, Fig. 3 presents the prediction results obtained using these two averaging methods.
The following useful information can be obtained from Fig. 3: First, the distribution of the prediction density obtained by the two averaging methods nearly coincide,which allows us to conveniently describe the prediction result without considering which method to use. Second, the prediction results of the single minimum error network are not better than the results obtained by averaging. Because the effect of randomness is significantly reduced by averaging the results, the following work in this study will use the averaging method to obtain the prediction results.
Fig. 2 (Color online) Charge density distributions obtained by FNNI3 and FNN-I4. The first row contains nuclei selected from the training set and the remaining nuclei are those selected from the validation set. The density distribution determined by the experimental parameters are indicated by solid black lines. The density distributions obtained by the FNN-I3 and FNN-I4 methods are indicated by the blue hatched and red hatched regions, and their mean predicted values are indicated by the blue-dashed and red-dotted lines, respectively.
Therefore, it is necessary to further explore the differences between the two averaging methods. Figure 4 presents the average curves and error bands of the charge density distributions obtained by the two methods for certain nuclei,and the selection of these nuclei is the same as that shown in Fig. 2. Apparently, the error bands obtained by averaging the density curves are narrower than those obtained by the other methods, although the average curves of the two are significantly coincident. This is because even a small change in the parameter has a large impact on the density distribution;thus, the uncertainty of the parameter is amplified by mapping. Therefore, the predicted charge density distribution and error band are obtained by the average density curves later in this study.
As the interpolation ability of FNN-I4 is verified, all the sets are adopted as learning sets to assess the predictive power of the neural network. Because the second and fourth moments of the charge density distribution ( 〈r2〉 and〈r4〉 )are important for understanding the nuclear structure (for example, using statistical correlation analysis, the diffraction radiusRand surface thicknessσare demonstrated to be sufficiently determined by〈r2〉 and〈r4〉 , especially for heavy nuclei [90, 91]), the learning effect of FNN-I4 can be evaluated using〈r2〉 and〈r4〉.
A comparison of the FNN-I4 predicted and experimental values of〈r2〉 and〈r4〉 are shown in Fig. 5a, b, respectively.Because the neural network directly outputs the parameters of the 2pF model, the predicted values of the second and fourth moments are calculated approximatively with a high precision as follows [92]:
Fig. 3 (Color online) The charge density distributions from different methods. The charge distributions obtained by using the experimental parameters are denoted by the solid black lines. The distributions obtained by using the mean predicted density and mean predicted parameter are indicated by the blue and red solid lines, respectively.The distributions obtained by the parameters that minimize (maximize) the loss function are denoted by the orange (purple)-dashed lines. The selection of nuclei is the same as that shown in Fig. 2
As shown in Fig. 5, the FNN results agree exceptionally well with the experimental values, especially for the lightmedium mass nuclei. The rms deviations (σ) between the experimental charge radii and the results of the FNN for〈r2〉and〈r4〉 are 0.7045 fm2and 42.83 fm4, respectively. Incidentally, the RMS deviation of the charge radii is 0.7041 fm .The large error between the predicted and experimental results for heavier nuclei is acceptable because there are few heavier nuclei.
The distributions of the learning and prediction sets are shown in Fig. 7, where the solid blue square represents the learning set, with a total of 86 nuclei. The solid red squares represent the prediction set, which has 284 nuclei. The nuclei in the prediction set are selected as follows: considering the nuclei in the stable nuclei as the endpoints, this isotope chain is filled, and the filled nuclides are included in the prediction set; the nuclei that are included in the learning set were eliminated. In Fig. 6, the charge radii of the prediction set nuclei, obtained from the FNN-I4 approach,are compared with the experimental values. The data of the experimental values are obtained from Ref. [93]. Among the 284 nuclei in the prediction set selected in this study, 230 have an experimental charge radii.
Figure 6 demonstrates that nearly all of the points are close to the liney=x, which indicates that the FNN-I4 predictions are significantly close to the experimental values,with the rms deviation (σ) between the experimental charge radii and the result of the FNN is 0.07693 fm . Certain nuclei significantly deviate from the experimental values; they are either light nuclei near the oxygen isotope chain or heavy nuclei withA>208 . These nuclei may have different physical properties that are not sufficiently reproduced by the current learning set. With the increase in the experimental data available for learning, the prediction accuracy of FNN-I4 in these regions may be improved. In summary, the FNN-I4 method presents a good prediction accuracy. The deviations of most predicted nuclei are less than 0.1 fm , which indicates that the FNN-I4 method is reliable for predicting the nuclear charge distribution. Therefore, the interpolation ability of the FNN-I4 method is also significantly promising.
Fig. 4 (Color online) The charge density distributions and error bands obtained by the average density distributions (Den) and the average parameters (Par) of multiple predictions, respectively, are indicated by the blue-dashed lines (blue hatched regions) and red-dotted lines(red hatched regions). The density curves determined by the experimental parameters are indicated by the solid black lines. The selection of nuclei is the same as that shown in Fig. 2
We also attempted to use the FNN-I4 method to predict the charge density distribution of calcium isotopes and obtained the charge radius from it. The experimental values of the charge radius are obtained from Ref.[93]. For the entire calcium isotope chain, only40Ca is included in the learning set, with the other nuclei serving as the prediction set. A few of the results are presented in Fig. 8. The uncertainty of the charge density in the nucleus center increases when it is distant from the nucleus of the learning set; therefore, only the prediction results near40Ca can be regarded as reliable, limiting the extrapolation ability of the FNN-I4 method. However, most of the predicted values are in good agreement with the experimental values, except that it is difficult to predict a sudden drop in the charge radius around48Ca .Considering that there is no corresponding situation in the existing learning set, adding learning samples can help the FNN approach in extracting physical information, which is helpful and necessary to improve the prediction accuracy.
In summary, we have employed the feedforward neural network approach to predict the nuclear charge density distribution, and the results obtained from the FNN are clearly demonstrated to be in good agreement with the experimental data. The nuclear 2pF model parameters obtained from Refs.[19—21] are studied. By adding the input variable ofA1∕3in the input layer, the machine learning method can accurately describe the 2pF model parameters of the nuclei (in the case of randomly dividing the dataset, the deviation in the training set is 0.02408 fm2for parametercand 0.00115 fm2for parameterz), and presents exceptional results in the validation set ( 0.02286 fm2for parametercand 0.00188 fm2for parameterz), which verifies the extrapolation ability of the FNN. Then, without any experimental values, the charge density distributions (described by the 2pF parameters) of the 284 nuclei are calculated using the FNN method. In addition, the density distribution of the calcium isotopes and the corresponding charge radii are obtained using the FNN method.
Fig. 5 (Color online) The differences between the experimental second (a) and fourth (b) moments of the charge density distribution and FNN training results (red dot)
Fig. 6 (Color online) Comparison of predicted rms charge radii(Rpre )and experimental values ( Rexp ) for nuclei with available data
Thus far, experimental data regarding the ground-state density distribution of spherical nuclei remains limited.Compared to traditional theoretical methods, the neural network method not only reduces the complexity of research and avoids complex multi-body problems, but also has a great prediction ability with less computational cost. In future, as the data available for neural networks to learn increase, the prediction ability will also be improved. Directly learning density distributions rather than using the model parameters may further improve the prediction.
Fig. 7 (Color online) learning set and prediction set in the nuclear chart. The learning sets are indicated by the solid blue square, including all 86 nuclei,and the prediction sets are indicated by the solid red square with a total of 284 nuclei. The two blue-dashed lines indicate the proton and neutron driplines; the numbers represent magic numbers
Fig. 8 (Color online) Charge density distributions of 38−49 Ca obtained by the FNN-I4 method. The predicted charge radius ( Rpre ) and experimental values ( Rexp ) of each nucleus are also shown in the figure; the values in the brackets after the Rpre indicate uncertainty. Those without experimental values are replaced by “—”
Declarations
Author contributionsAll authors contributed to the study conception and design. Material preparation, data collection and analysis were performed by Tian-Shuai Shang, Jian Li and Zhong-Ming Niu. The first draft of the manuscript was written by Tian-Shuai Shang and all authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Nuclear Science and Techniques2022年12期