Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.
Scientific Reports volume 14, Article number: 4681 (2024 ) Cite this article rotary joint for steam
The preparation process and composition design of heavy-section ductile iron are the key factors affecting its fracture toughness. These factors are challenging to address due to the long casting cycle, high cost and complex influencing factors of this type of iron. In this paper, 18 cubic physical simulation test blocks with 400 mm wall thickness were prepared by adjusting the C, Si and Mn contents in heavy-section ductile iron using a homemade physical simulation casting system. Four locations with different cooling rates were selected for each specimen, and 72 specimens with different compositions and cooling times of the heavy-section ductile iron were prepared. Six machine learning-based heavy-section ductile iron fracture toughness predictive models were constructed based on measured data with the C content, Si content, Mn content and cooling rate as input data and the fracture toughness as the output data. The experimental results showed that the constructed bagging model has high accuracy in predicting the fracture toughness of heavy-section ductile iron, with a coefficient of coefficient (R2) of 0.9990 and a root mean square error (RMSE) of 0.2373.
Heavy-section ductile iron is a widely used cast iron material. Due to its good fracture toughness, tensile strength, plasticity and antiradiation properties, it is commonly used in harshly demanding fields such as nuclear power, wind power, storage and transportation of nuclear spent fuel and high-speed railroads1,2,3. In view of the increasingly complex working conditions of heavy-section ductile iron, its safety is the most important consideration in engineering applications, and fracture toughness is an important basis for safety assessment4. At present, only German, Japanese, and individual scholars have carried out fracture toughness mechanism research on nuclear spent fuel storage and transportation containers with a wall thickness of less than 150 mm5,6,7. Panneerselvam8 studied the low-temperature fracture toughness of heavy-section ductile iron containers with 150 mm wall thickness, and the results showed that an increase in pearlite content led to a decrease in fracture toughness. Singh9 found that the fracture surface of heavy-section ductile iron has a special metal oxide film, which greatly reduced the fracture toughness and elongation of the material. The film was observed was due to the long cooling time, which is obviously different from ordinary ductile iron. Song and Guo10,11 studied the influence of silicon content and nodulizers on heavy-section ductile cast iron, and found that lower silicon content can reduce the performance of chunky graphite iron, especially in its capacity to demonstrate good impact performance at low temperatures.
At present, research on the fracture toughness mechanism of heavy-section ductile iron is relatively limited, and the chemical compositions of the main components of heavy-section ductile iron influence and interfere with each other, resulting in high preparation costs and long design cycles, and systematic research on prediction of heavy-section ductile iron fracture toughness is scarce, which seriously affects the practical application of heavy-section ductile iron.
Turing Award winner James Gray classifies scientific research into four types of paradigms: experimental induction, model derivation, simulation and data-intensive scientific discovery. As part of the fourth paradigm of scientific research, ML is a multidisciplinary field that covers numerous disciplines. Machine learning is mainly the study of how computers mimic human learning behavior, acquire new knowledge or experience, and reorganize existing knowledge structures to improve their performance, and ML is also widely used in other fields of scientific research12.
Machine learning has been widely used in materials design. Chiniforush13 achieved phase prediction of novel high-entropy alloys based on the random forest classifier model. Anand12 made a breakthrough in topological feature engineering for perovskite material design by using a machine learning model. Huo14 utilized a semisupervised machine learning approach to unlock a large amount of inorganic materials in literature synthesis information and processed it into a standardized machine-readable database. Schmidt15 used a machine learning model to screen out materials with high catalytic activity and predicted and optimized the catalytic efficiency of the materials by establishing a relationship between the material surface structure and the rate of catalysis reaction. However, the prediction of the fracture toughness of heavy-section ductile iron has not been reported.
In actual production, the five major elements for the preparation of heavy-section ductile iron are mainly carbon (C), silicon (Si), manganese (Mn), sulfur (S), and phosphorus (P). These elements and cooling rate have an important influence on the properties of heavy-section ductile iron24,25,26,27,28,29. The following is also the basis for the selection of chemical composition range for the preparation of samples in this paper.
C is the basic element of heavy-section ductile iron and helps in graphitization, and the general content of Carbon is controlled between 3.5 and 3.9 wt%. Too high carbon content leads to graphite floating, too low leads to low spheroidization rate and reduced mechanical properties.
Si is strong graphitization element, which can effectively reduce the white iron tendency, increase the amount of ferrite, and improve the roundness of graphite ball. However, silicon will also increase the ductile–brittle transition temperature and reduce the impact toughness. Therefore, the silicon content should not be too high, generally controlled between 1.7 and 3.8 wt%.
The main role of Mn is to increase the stability of pearlite and promote the formation of carbides, these carbides segregate at grain boundaries, which greatly reduces the toughness of ductile iron. The preparation of heavy-section ductile iron generally requires that it has a ferrite matrix to obtain high fracture toughness. However, due to the long cooling time of the core, it is easy to generate a large amount of pearlite and affect its toughness. Generally, the Mn content is limited to less than 0.5 wt%.
In the preparation of heavy-section ductile iron, the content of S and P is often controlled as low as possible. Excessive S content will lead to hot brittleness and excessive P content will lead to cold brittleness of ductile iron.
For heavy-section ductile iron, due to the increased wall thickness, casting solidification is slow, spheroidal fading, coarse graphite, chunky graphite, elemental segregation and other defects easily occur, and the microstructure and fracture toughness of heavy-section ductile iron are significantly different from those of normal ductile iron. To obtain high fracture toughness heavy-section ductile iron, water cooling is generally used to increase the cooling solidification rate. However, because the actual production of the heavy-section ductile iron wall thickness is different, even with the use of water cooling, the solidification time of the core test block is often more than 150 min, and the method cannot unlimitedly reduce its cooling rate to increase its fracture toughness.
In order to explore the relationship between the four factors (C content, Si content, Mn content, cooling rate), fracture toughness and the microstructure, samples with different (C content, Si content, Mn content, cooling rate) were selected for research, as shown in Fig. 1. Figure 1 shows the effects of C, Si, Mn and cooling rate (part of the measured data is selected) on the microstructure of heavy-section ductile iron. Figure 1a,b have different C contents, which are 3.3 wt% and 3.6 wt%, respectively, and the other components are the same. Figure 1c,d have different Si contents of 1.9 wt% and 2.3 wt%, respectively, and the other components are the same. Figure 1e,f have different Mn contents, which are 0.1 wt% and 0.7 wt%, respectively, and the other components are the same. Figure 1g,h have different cooling rates, 145 min and 265 min, respectively, and the other components are the same.
Effect of C, Si, Mn and cooling time on micro-structure of heavy-section ductile iron (part of the measured data) (a) C contents is 3.3 wt%, (b) C contents is 3.6 wt%, (c) Si contents is 1.9 wt%, (d) Si contents is 2.3 wt%, (e) Mn contents is 0.1 wt%, (f) Mn contents is 0.7 wt%, (g) Cooling time (min) 145, (h) Cooling time (min) 265.
It can be seen from Fig. 1a,b that an appropriate increase in C content can increase the promotion of graphite nucleation and increase the number of graphite spheres. When the Si content reaches 2.3 wt%, the number of graphite spheres increases and the roundness increases, as shown in Fig. 1c,d. The increase of Mn content leads to a significant increase in the number of pearlite in the matrix structure, and a large amount of chunky graphite is produced, as shown in Fig. 1e,f. When the cooling rate of heavy section ductile iron is accelerated, the number of graphite balls is greatly increased and the graphite morphology is improved, as shown in Fig. 1g,h.
Figure 2 is the fracture morphology of fracture toughness test of the measured data in this paper (part of the measured data is selected), the composition of each sample is the same as Fig. 1.
The fracture toughness fracture morphology of heavy-section ductile iron (a) C contents is 3.3 wt%, (b) C contents is 3.6 wt%, (c) Si contents is 1.9 wt%, (d) Si contents is 2.3 wt%, (e) Mn contents is 0.1 wt%, (f) Mn contents is 0.7 wt%, (g) Cooling time (min) 145, (h) Cooling time (min) 265.
It can be seen from Fig. 2a,b that an appropriate increase in C content can increase the number of graphite spheres, and a large number of graphite spheres play a role in releasing stress concentration when subjected to external forces, and the heavy-section ductile iron has more dimples and is more inclined to ductile fracture. It can be seen from Fig. 2c,d that when the Si content reaches 2.3 wt%, the increase in the number of graphite spheres and the increase in roundness also lead to the development of the heavy-section ductile iron from ductile–brittle mixed fracture to ductile fracture. The increase of Mn content leads to a significant increase in the number of pearlite in the matrix structure, and a large amount of chunky graphite is produced, which leads to a large number of cleavage planes and river patterns on the fracture surface, and the fracture morphology changes from ductile–brittle mixed fracture to brittle fracture, As shown in Fig. 2e,f. When the cooling rate of heavy-section ductile iron is accelerated, a large number of small and round graphite balls make the fracture morphology show a typical ductile fracture, As shown in Fig. 2g,h.
Figure 3 shows the measured data in this paper. The influence of C, Si, Mn and cooling speed (selected part of the measured data) on the fracture toughness of heavy-section ductile iron was investigated. Figure 3 shows that the chemical elements C, Si, Mn on the fracture toughness of heavy-section ductile iron present a nonlinear relationship, and only the cooling rate on the fracture toughness effect shows a linear relationship. With increasing cooling rate, the fracture toughness increases. This leads to the preparation of high fracture toughness heavy-section ductile iron parts in actual production, and a large number of tests need to be carried out to explore the combination of a reasonable chemical composition and forced cooling method.
Effect of C, Si, Mn and cooling time on the fracture toughness of heavy section ductile iron (part of the measured data).
In this paper, a total of 18 large heavy-section ductile iron test blocks with different chemical compositions were cast. Benxi Q12 pig iron, 75% silicon iron, steel grade 45 and graphite powder were melted in a medium-frequency induction furnace, and a spheroidizing treatment was applied using a Ce–Mg–Si nodulizing agent. The molten metal was poured into sand using furan resin sand to obtain heavy-section ductile iron cubic test blocks with a size of 400 mm × 400 mm × 400 mm. The chemical composition of nodulizing agent is shown in Table 1, and 18 heavy-section ductile iron test blocks with different C, Si, and Mn contents were labeled Casting 1, Casting 2, Casting 3, and Casting 18, and their chemical compositions are listed in Table 2.
Four positions in castings chosen for temperature measurement and specimen collection.
Temperature measurement process of pouring and the cooling time of the four positions in castings (a) Casting process, (b) temperature measurement process, (c) cooling time of four positions in castings.
In this paper, the fracture toughness prediction of heavy-section ductile iron was investigated, the chemical elements C, Si, Mn and cooling time were the four factors used as input data for machine learning, and the fracture toughness was used as the output data to establish a machine learning prediction model. The total number of samples was 72, and all sample data were measured data.
The fracture toughness was tested at room temperature using an MTS 809 electrohydraulic servo material testing machine according to the standard GB/T4161-2007. The dimensions of the fracture toughness compact tensile (CT) specimens are shown in Fig. 6.
Dimensions of the fracture toughness specimen in mm.
In this paper, six machine learning methods were selected to study the influence of C, Si, Mn and the cooling rate on the fracture toughness of heavy-section ductile iron, and corresponding machine learning model of regression prediction of fracture toughness of heavy-section ductile iron was constructed.
XGBoost is a parallel regression tree model based on the idea of boosting. Boosting refers to the weighted summation of a number of existing classifiers to obtain the final classifier. The XGBoost model was improved on the basis of the gradient descent decision tree (GBDT) by Chen30. In this model, each nonleaf node represents a feature, and leaf nodes represent a kind of label or decision result. When applying the decision tree, the samples to be predicted are examined for judgment conditions according to the feature values corresponding to the nodes to determine the next node position and judgment conditions and iteratively until a definite decision result is obtained.
Since the decision tree structure is simple and logical, it is prone to overfitting. Therefore, a random forest approach to reduce overfitting has been derived. The random forest constructs multiple random training sets and trains a weak learner decision tree based on them. Then, the decision trees are integrated to obtain a strong learner decision tree with superior performance to avoid overfitting31. Since the integrated decision trees are independent of each other and without feedback, a gradient descent decision tree method was derived.
Each tree in the GBDT (gradient-boosted decision tree) is fitted using the residuals (i.e., error-free observations) of the previous tree, and the final result is determined by the sum of the results of all trees. However, this also makes it impossible to perform parallel operations. XGBoost is further extended based on the GBDT method by presorting and saving the training data before the model is trained and using the sorted data in the iteration process32. This model calculates the feature gain at the time of new node selection and selects the node with the larger value to split and form the next layer of child nodes. Due to the use of prearranged data, XGBoost can split multiple features at the same time, thus realizing parallel operation and saving model training time. At the same time, its objective function not only includes the common loss function but also adds a regularization term and uses the column method in random forest, which reduces overfitting and accelerates the speed of parallel computation. The most important core issue of the XGBoost algorithm is the objective function that combines the evaluation error of decision trees:
In the equation, Loss is the loss function of the deviation between the actual value and the predicted value, \(y_{j}\) is the actual value of the predicted data, \(\hat{y}_{i}\) is the result of the previous decision tree algorithm, x is the input data, thus fitting the algorithm results of multiple decision trees, while \(f_{m}\) is the approximation function used by the decision tree. Meanwhile, \(\gamma\) is the regularization term for improving the penalty coefficient used by the decision tree, and the basic decision tree approximation function is:
In this paper, the XGBoost model parameters are searched by the random search method, and fivefold cross validation is used to evaluate the model performance to finally arrive at the best model. The model graph is shown in Fig. 7.
The regression algorithm is support vector regression or SVR. SVR is a supervised learning algorithm used to predict discrete values. Support vector regression uses the same principles as those in SVM. The basic idea is to map the data into a high-dimensional space and find the optimal hyperplane for regression and thus the line of best fit. Compared with traditional regression algorithms, SVR not only considers the degree of data fit but also the generalizability of the model. Thus, SVR can effectively deal with high-dimensional data and nonlinear data33. In the regression task, it is necessary to make the interval between the sample points that are farthest away from each other by the hyperplane the largest. That is, the SVR is given a restriction on the interval, and the deviation of the regression model f (x) from y must be \(\le \varepsilon\) for all sample points. This range is referred to as the deviation of the ε pipeline.
The SVR optimization problem can be expressed mathematically as34:
Given samples \(D = \{ (x_{1} ,y_{1} )(x_{2} ,y_{2} ),...,(x_{n} ,y_{n} )\} ,{\text{y}}i \in R\) , f (x) is obtained such that it is as close as possible to y, and w and b are the parameters to be determined. The loss is zero when f (x) and y are identical. Support vector regression assumes that there is at most ε deviation between f (x) that can be tolerated. The loss is calculated when and only when the absolute value of the difference between f (x) and y is greater than ε, which is equivalent to the construction of a spacing band with a width of 2ε centered around f (x). The training samples are considered to be correctly predicted if they fall into this spacing band, as shown in Fig. 8.
Structure of the SVR model.
The MLP regressor is a supervised learning algorithm. Figure 9 shows the MLP model with only 1 hidden layer; the left side is the input layer, and the right side is the output layer35.
The MLP is also known as the multilayer perceptron. In addition to the input and output layers, it can have more than one hidden layer in the middle. A linearly divisible data problem can be solved if there is no hidden layer36. The layers of the multilayer perceptron shown in Fig. 9 are fully connected to each other. Therefore, it is possible to solve the problem without any hidden layer. The bottom layer of the multilayer perceptron is the input layer, the middle layer is the hidden layer, and the last layer is the output layer. The input layer is a 4-dimensional vector, and the output is 4 neurons37. The neurons in the hidden layer are fully connected to the input layer. It is assumed that the input layer is represented by the vector Xi, and the output of the hidden layer is f (WkjX + b1), Wji is the weight (also known as the connection coefficient), b1 is the bias (in the design model in Fig. 9, the bias is 0 by default).The function f can be the commonly used sigmoid function or tanh function. Finally, the output layer is connected to the hidden layer through a sigmoid function or tanh function. This connection is analogous to a multicategory logistic regression, that is, softmax regression. Therefore, the output of the output layer is y = softmax (WkjXj + b2), where Xj represents the hidden layer output f (WkjXj + b2), and b2 is the bias.
The Gaussian process model is a statistical tool for constructing stochastic processes and is widely used in the fields of machine learning, statistics and information processing. The Gaussian process model is based on the principles of probability theory. By modeling and predicting observed data, this model can provide estimates of the probability distribution of unknown data points38.
In the Gaussian process model, it is assumed that the stochastic process under study obeys a multivariate Gaussian distribution for any set of inputs. Each point in the input space is usually projected to a random variable in the output space, and the joint assignment of these random variables constitutes the probabilistic model of the Gaussian process39.
A Gaussian process model can be described by two underlying components: the mean value function and the variance function (also known as the kernel function). The mean value function represents modeling the overall dynamics of the stochastic overprocess, while the covariance function describes the correlation or similarity between different points40.
Given a set of input points and corresponding observations, predictions can be made using a Gaussian process model. The result of the prediction is a conditional probability distribution over the unknown data points, which includes a predicted mean and a predicted uncertainty (variance). This estimate of uncertainty is quantified by the covariance function, which portrays the correlation between the input points, reflecting the uncertainty of the expected measurements15. Its mathematical expression is as follows:
Here, x is the mean, and y is the variance. The expression is as follows:
Considering that the general observations of the function are noisy, there is:
where \(\varepsilon \sim N(0,\sigma_{n}^{2} )\) is the white noise with a variance of \(\sigma_{n}^{2}\) . The prior distribution of y can be expressed as:
At the test dataset \({\varvec{x}}^{ * }\) , the joint prior distribution of the observation set y and the prediction set ŷ can be expressed as:
According to the above equation, the posterior distribution \({\text{p}}({\hat{\mathbf{y}}}|{\varvec{x}},{\varvec{y}},{\varvec{x}}^{ * } )\) can be obtained:
where the output of the model prediction is the predicted mean \({\hat{\mathbf{y}}}^{ * }\) , and the uncertainty of the prediction model is reflected by the prediction covariance \({\text{cov}}(\hat{y}^{ * } )\) . The expression is as follows:
The standard Gaussian process model can be represented as shown in Fig. 10.
Diagram of the standard Gaussian process model.
Bagging regression is an integrated learning method that generates multiple subsets of training data by randomly sampling the training data with replacement and then uses these subsets to train multiple base learning devices41,42. The main idea of bagging regression is to improve the generalizability of the model by reducing the variance. The specific process is to generate multiple subsets from the training set X by random sampling in a relaxed manner and use each subset to train a base learner. Each base learner is a regression model, such as decision tree regression or linear regression43. The main idea of bagging regression is to improve the generalizability of the model by reducing the variance. After training, the prediction results of the base learners are averaged or weighted average, and the final prediction results are obtained Y. Bagging regression reduces the variance of the model and improves the generalizability of the model. Parallelization is possible at the same time, accelerating model training. The process is shown in Fig. 11.
Schematic diagram of the bagging algorithm.
Let the expectation of the single model be \(\mu\) ; then, the expectation of the bagging regression is:
n is the number of base learners, \(X_{i}\) is the expectation of the i base learner. This expectation value represents the expected performance of the Bagging regression model on the entire training dataset, helping us evaluate the overall performance of the Bagging regression model.
The random forest is a more optimized combinatorial forecasting model proposed by professor Breiman in 2001 based on the idea of bagging, which is a new extension of randomness44,45. The random forest algorithm utilizes bootstrap sampling to construct different tree models with random samples. Then, the selection of the best node of each tree model is changed so that the variable node of each tree also has randomness, which in turn generates a number of regression trees with high prediction accuracy and uncorrelated regression trees46. The predictions of all the trees are average to vote, representing the prediction of this regression model. The RF model structure is shown in Fig. 12.
Let {\(\left\{ {h\left( {x,\theta_{t} } \right)} \right\}\) , t = 1,2,…,T} be a T regression tree in a random forest, where \(\theta_{{\text{t}}}\) is a random variable obeying an independent homogeneous distribution and x is a dependent variable. Then, the regression result can be expressed as:
The model is trained on measured fracture toughness data of large ductile cast iron. The hardware environment of the simulation platform for this experiment is a CPU and a GPU. The CPU is an Intel (R) Core i7-10700 @ 2.90 GHz, and the GPU is an NVIDIA GeForce GTX 1660 SUPER with 64 GB of RAM. The simulation software environment uses PyCharm, Python version 3.8, and the Sklearn, NumPy, pandas and matplotlib libraries.
Among them, the R2 value is in the range of (0, 1). The closer to 1 the value is, the better the prediction result. In contrast, the closer to 0 the value is, the worse the prediction result. The RMSE value is in the range of (zero, + ∞). A value closer to 0 indicates a smaller prediction error.
Genetic algorithm is a heuristic search and optimization technique inspired by the processes of biological evolution in nature. It simulates the processes of biological evolution, such as selection, crossover, and mutation, to find the optimal solution or better solution to the problem. It is widely used in the field of machine learning. In this study, the samples are trained using the fifty percent cross-validation method, and the hyperparameter optimization method is the genetic algorithm (GA). XGBoost has numerous hyperparameters that need to be manually set. In this paper, we selected several common hyperparameters: n_estimators (number of trees), max_depth (maximum tree depth), learning rate, and subsample. For the SVR model, we chose C (penalty parameter), gamma (kernel coefficient), and kernel type. The MLP Regressor model selected alpha (regularization parameter), max_iter (maximum number of iterations), and solver for optimizer selection. Gaussian Process Regression model parameters included alpha (value added to the diagonal of the kernel matrix during model fitting) and kernel type. For the Bagging model, we chose n_estimators (number of base estimators), max_samples (number of samples to train base estimators), max_features (number of features to train base estimators), and bootstrap (determines the sampling method for the sample subset). The Random Forest model selected n_estimators (number of trees) and maxdepth (maximum tree depth). The specific model coefficients and optimized coefficients are shown in Table 3.
Figure 13 illustrates the comparison of the goodness of fit of six machine learning models optimized through genetic algorithms. As shown in Fig. 13, the optimized XGBoost, SVR, MLP, and Random Forest models exhibit poor fitting of predicted values to actual values, while the optimized Bagging and Gaussian process models show a good fit between predicted and actual values, approaching a single line.
Comparison of the degree of fit of the models (a) XGBoost, (b) SVR, (c) MLP, (d) Gaussian process regression, (e) Bagging, (f) Random forest.
To further compare the accuracy of the Gaussian process and bagging models in predicting the fracture toughness of heavy-section ductile iron, the RMSE and R2 are selected as the two metrics to measure the reliability and accuracy of the model predictions. As shown in Fig. 14, the R2 values of the XGBoost, SVR, MLP regressor, Gaussian process and random forest models are 0.8662, 0.8901, 0.5942, 0.99 and 0.63, respectively, which are lower than the R2 value of bagging (0.9990). The RMSEs of the XGBoost, SVR, MLP regressor, Gaussian process and random forest models are 3.085, 0.661, 4.5467, 0.3937 and 5.21, respectively, which are higher than the RMSE of 0.2373 of the bagging model. Therefore, the bagging model is better for predicting the fracture toughness of heavy-section ductile iron.
Evaluation indicators for each model.
The bagging model optimized by the genetic algorithm is applied to 72 fracture toughness specimens of heavy-section ductile iron with different compositions and cooling times constructed in this paper. The experimental results are shown in Fig. 15a. The results indicate that the projected value and the true value data points basically coincide with each other, and the prediction effect is accurate. Figure 15b shows the absolute value of the prediction error. The figure shows that the prediction error is basically less than 0.6, and the maximum does not exceed 0.8.
Bagging model prediction results (a) Comparison of the true and predicted values, (b) absolute error.
Bagging is an ensemble learning method that works by constructing multiple weak learners (typically decision trees) and combining their results to improve the overall model performance. On the other hand, genetic algorithm is an optimization algorithm that simulates the biological evolution process, searching for the optimal solution to a problem by mimicking evolutionary mechanisms such as natural selection, crossover, and mutation. Through genetic algorithm optimization, the Bagging model can better adapt to different data patterns, improve prediction accuracy for unknown data, and particularly demonstrate significant advantages for materials with complex structures such as thick-sectioned ductile cast iron.
In this paper, six machine-learning models optimized by the genetic algorithm are applied to 72 heavy-section ductile iron specimens with different compositions and cooling times. The experimental results show that the optimized bagging model has the best prediction effect, with an R2 of 0.9990 and an RMSE of 0.2373.
The general C content of heavy-section ductile iron is controlled between 3.5 and 3.9 wt%, appropriate carbon content can make cast iron have good fluidity and lubricity, which is convenient for filling mold cavity. However, if the C content is too high to exceed 3.9 wt%, the plasticity and toughness of heavy-section ductile iron will be seriously reduced, and the ductile iron is prone to crack and fracture, the thermal brittleness increases47. The general Si content of heavy-section ductile iron is controlled between 1.7 and 3.8 wt%, the hardness, tensile strength and yield strength of heavy-section ductile iron can be improved by adding appropriate amount of silicon. However, the ductile–brittle transition temperature of ductile iron can be significantly increased and the plastic and toughness of heavy-section ductile iron will be reduced when the silicon content is too high to exceed 3.8 wt%48. The preparation of heavy-section ductile iron generally requires its ferrite matrix to obtain high fracture toughness. Due to the long cooling time in the core of heavy-section ductile iron, it is easy to generate a large amount of pearlite, and its Mn content needs to be strictly controlled, generally controlled within 0.5 wt%49.
In this paper, in order to explore the performance prediction of fracture toughness of heavy-section ductile iron products, the content range of C, Si and Mn of heavy-section ductile iron is expanded in the additional test, as shown in Table 4.
As shown in Table 4, additional experiments were conducted to prepare four heavy-section ductile iron specimens with varying carbon (C) content, silicon (Si) content, and manganese (Mn) content. A total of 16 samples were cut from the edge to the core of each specimen, and the cooling time was the same as that of the previous samples. The basic parameters for each model were the same as those in Table 3, and the same methods were employed.
Figure 16 compares the fitting degree of six machine learning models. From the figure, it can be seen that the fitting effect of the Bagging model optimized by genetic algorithm and the Gaussian process regression model is still better, approaching a straight line.
Comparison of the fitting degree of each model. (a) XGBoost, (b) SVR, (c) MLP, (d) Gaussian process regression, (e) Bagging, (f) Random forest.
To further compare the accuracy of Gaussian Process and Bagging models in predicting the fracture toughness of heavy-section ductile iron, this study selected the root mean square error (RMSE) and coefficient of determination (R2) as two indicators to measure the reliability and accuracy of the model predictions. As shown in Fig. 17, the R2 values for XGBoost, SVR, MLP Regressor, Gaussian Process, and Random Forest models are 0.9061, 0.9435, 0.7105, 0.9818, and 0.7582, respectively, all lower than Bagging 0.9873. Additionally, the RMSE values for these models are 0.6936, 1.0537, 2.38, 0.5938, and 2.17, all higher than Bagging’s 0.4993. The results indicate that after expanding the component range of heavy-section ductile iron products, the obtained results still meet the requirements for practical applications.
Evaluation indicators for each model.
In this paper, four factors affecting the fracture toughness of heavy-section ductile iron, such as C content, Si content, Mn content and cooling rate, are discussed. By using the isothermal section method, 18 cubic physical simulation specimens with different compositions and wall thicknesses of 400 mm were cast, and 72 specimens were prepared for microstructure observation and fracture toughness testing. In addition, the relationship between the above four factors and the microstructure and fracture toughness of heavy-section ductile iron was discussed.
Aiming at the problems of a long preparation cycle, high R&D cost, and many nonlinear influences for high fracture toughness heavy-section ductile iron, a machine-learning-based fracture toughness prediction model for thick and large section ductile iron was established. The C content, Si content, Mn content and cooling rate are input, and the fracture toughness is output.
Compared with the XGBoost, SVM, MLP regressor, Gaussian process, random forest and other models, the bagging model has the best prediction effect, followed by the Gaussian process, with the R2 values reaching 0.9990 and 0.99, respectively, and RMSE values of 0.2373 and 0.3937, respectively. These models can meet the design requirements of high fracture toughness heavy-section ductile iron for nuclear spent fuel storage and transportation containers and bases of wind power.
The datasets used and analyzed during the current study available from the corresponding author on reasonable request.
Padmakumar, M. & Arunachalam, M. Analyzing the effect of cutting parameters and tool nose radius on forces, machining power and tool life in face milling of ductile iron and validation using finite element analysis. Eng. Res. Express. 2, 1–13 (2020).
Yang, P. H. et al. Experimental and ab initio study of the influence of a compound modifier on carbidic ductile iron. Metall. Res. Technol. 116, 306–311 (2019).
Cheng, H. Q. et al. Effect of Cr content on microstructure and mechanical properties of carbidic austempered ductile iron. Mater. Test. 60, 31–39 (2018).
Article ADS CAS Google Scholar
Chiniforush, E. A., Yazdani, S. & Nadiran, V. The influence of chill thickness and austempering temperature on dry sliding wear behaviour of a Cu–Ni carbidic austempered ductile iron (CADI). Kovove Mater. 56, 213–221 (2018).
Kusumoto, K. et al. Abrasive wear characteristics of Fe–2C–5Cr–5Mo–5W–5Nb multicomponent white castiron. Wear 3, 22–29 (2017).
Foglio, E., Lusuardi, D. & Pola, A. Fatigue design of heavy section ductile irons: Influence of chunky graphite. Mater. Des. 111, 353–361 (2016).
Ceschini, L., Morri, A. & Morri, A. E. Microstructure and mechanical properties of heavy section ductile iron castings: Experimental and numerical evaluation of effects of cooling rates. Taylor Francis 28, 294–316 (2015).
Panneerselvam, S. et al. An investigation on the stability of austenite in austempered ductile cast iron (ADI). J. Mater. Sci. Eng. A 626, 237–246 (2016).
Singh, S. et al. Phase prediction and experimental realisation of a new high entropy alloy using machine learning. Sci. Rep. 13, 48–53 (2023).
Liang, S., Erjun, G., LiPing, W. & Dongrong, L. Effects of silicon on mechanical properties and fracture toughness of heavy-section ductile cast iron. Metals-Basel 5, 150–161 (2015).
Erjun, G., Liang, S. & Liping, W. Effect of Ce–Mg–Si and Y-Mg–Si nodulizers on the microstructures and mechanical properties of heavy section ductile iron. J. Rare Earths 32, 734–743 (2014).
Anand, D. V. et al. Topological feature engineering for machine learning based halide perovskite materials design. NPJ Comput. Mater. 8, 203–209 (2022).
Chiniforush, E. A., Yazdani, S. & Nadiran, V. The influence of chill thickness and austempering temperature on dry sliding wear behaviour of a Cu–Ni carbidic austempered ductile iron (CADI). Kovove Mater. 56, 213–221 (2021).
Huo, H. et al. Semi-supervised machine-learning classification of materials synthesis procedures. NPJ Comput. Mater. 5, 62–70 (2019).
Schmidt, J. et al. Recent advances and applications of machine learning in solid-state materials science. NPJ Comput. Mater. 5, 83–91 (2019).
Liu, C., Du, Y. & Wang, X. Comparison of the tribological behavior of quench-tempered ductile iron and austempered ductile iron with similar hardness. Wear 5, 146–175 (2023).
Borsato, T. & Fabrizi, A. Long solidification time effect on solution strengthened ferritic ductile iron fatigue properties. Int. J. Fatigue 145, 1–7 (2021).
Stewart, B. C. et al. Comparison study of ductile iron produced with Martian regolith harvested iron from ionic liquids and Bosch byproduct carbon for in-situ resource utilization versus commercially available 65-45-12 ductile iron. Adv. Space Res. 71, 2175–2185 (2023).
Article ADS CAS Google Scholar
Stewart, B. C. et al. Effects of nickel and manganese on ductile iron utilizing ionic liquid harvested iron and Bosch byproduct carbon. Acta Astronaut. 204, 175–185 (2023).
Article ADS CAS Google Scholar
KeRui, L., LiangLiang, Z. & WeiDong, Z. Research and optimization of casting process for 100-ton ductile iron nuclear spent fuel container. Foundry 69, 09–717 (2020).
LinFu, D., ShanZhi, R. & Rui, L. Study on simulated production test of spent fuel rare earth magnesium ductile iron container. Mod. Cast Iron. 4, 7–11 (1994).
LinFu, D., RenFeng, D. & ShuFan, C. Quality control of heavy-section ductile iron castings. Foundry 69, 7–30 (1997).
Guo, H., Zhao, J. Y. & Yin, J. H. Random forest and multilayer perceptron for predicting the dielectric loss of polyimide nanocomposite films. RSC Adv. 7, 30999–31008 (2017).
Article ADS CAS Google Scholar
Foglio, E. et al. Fatigue characterization and optimization of the production process of heavy section ductile iron castings. Int. J. Met. 11, 33–43 (2017).
Mourad, M. M. et al. Effect of processing parameters on the mechanical properties of heavy section ductile iron. J. Metall. 2015, 1 (2015).
Cai, Q. et al. Effect of elemental segregation on the microstructure and mechanical properties of heavy section compacted graphite iron. Int. J. Met. 17, 222–232 (2023).
Bauer, B. et al. Influence of chemical composition and cooling rate on chunky graphite formation in thick-walled ductile iron castings. Int. J. Met. 17, 050–2061 (2023).
Benedetti, M., Fontanari, V. & Lusuardi, D. Effect of graphite morphology on the fatigue and fracture resistance of ferritic ductile cast iron. Eng. Fract. Mech. 206, 427–441 (2019).
Wang, Y., Song, R. & Huang, L. The effect of retained austenite on the wear mechanism of bainitic ductile iron under impact load. J. Mater. Res. Technol. 1, 1665–1671 (2021).
Chen, T. & Guestrin, C. XGBoost: A scalable tree boosting system. Preprint at https://doi.org/10.48550/arXiv.1603.02754 (2016).
Kigo, S. N., Omondi, E. O. & Omolo, B. O. Assessing predictive performance of supervised machine learning algorithms for a diamond pricing model. Sci. Rep. 13, 17315 (2023).
Article ADS CAS PubMed PubMed Central Google Scholar
Domingo, L. et al. Binding affinity predictions with hybrid quantum-classical convolutional neural networks. Sci. Rep. 13, 17951 (2023).
Article ADS CAS PubMed PubMed Central Google Scholar
Matsuo, H. et al. Machine learning-based prediction of relapse in rheumatoid arthritis patients using data on ultrasound examination and blood test. Sci. Rep. 12, 7224 (2022).
Article ADS CAS PubMed PubMed Central Google Scholar
Nakajima, M. & Nemoto, T. Machine learning enabling prediction of the bond dissociation enthalpy of hypervalent iodine from SMILES. Sci. Rep. 11, 20207 (2021).
Article ADS CAS PubMed PubMed Central Google Scholar
Nasir, V. & Sassani, F. A review on deep learning in machining and tool monitoring: Methods, opportunities, and challenges. Int. J. Adv. Manuf. Technol. 115, 2683–2709 (2021).
Liu, Q., Shi, W. & Chen, Z. Rubber fatigue life prediction using a random forest method and nonlinear cumulative fatigue damage model. J. Appl. Polym. Sci. 137, 48519 (2020).
Pei, H. et al. A review of machine learning-based methods for predicting the remaining life of equipment. J. Mech. Eng. 55, 1–13 (2019).
Su, K. et al. Prediction of fatigue life and residual stress relaxation behavior of shot peening 25CrMo alloy based on neural network. Rare Metal Mater. Eng. 49, 2697–2705 (2020).
Zhou, T., Song, Z. & Sundmacher, K. Big data creates new opportunities for materials research: A review on methods and applications of machine learning for materials design. Engineering 5, 1017–1026 (2019).
Zhao, C. et al. Optimization of TC4 material process parameters based on neural network genetic algorithm for magnetic particle grinding. Surf. Technol. 49, 316–321 (2020).
Muraoka, K. et al. Linking synthesis and structure descriptors from a large collection of synthetic records of zeolite materials. Nat. Commun. 10, 44–59 (2019).
Smer-Barreto, V. et al. Discovery of senolytics using machine learning. Nat. Commun. 14, 34–45 (2023).
Ge, Q. et al. Modelling armed conflict risk under climate change with machine learning and time-series data. Nat. Commun. 13, 28–39 (2022).
Ramprasad, R. et al. Machine learning in materials informatics: Recent applications and prospects. NPJ Comput. Mater. 3, 54–62 (2017).
Guo, Z. et al. Fast and accurate machine learning prediction of phonon scattering rates and lattice thermal conductivity. NPJ Comput. Mater. 9, 95–99 (2023).
Jiang, Y. et al. Topological representations of crystalline compounds for the machine-learning prediction of materials properties. NPJ Comput. Mater. 7, 28–36 (2021).
Article ADS CAS PubMed PubMed Central Google Scholar
Riebisch, M., Pustal, B. & Bührig-Polaczek, A. Impact of carbide-promoting elements on the mechanical properties of solid-solution-strengthened ductile iron. Int. J. Met. 14, 365–374 (2020).
Nam, J. H., Lee, S. M. & Lee, S. H. Guaranteed soundness of heavy section spheroidal graphite cast iron based on a reliable C and Si ranges design. Met. Mater. Int. 29, 2151–2158 (2023).
Benedetti, M. et al. Multiaxial plain and notch fatigue strength of thick-walled ductile cast iron EN-GJS-600-3: Combining multiaxial fatigue criteria, theory of critical distances, and defect sensitivity. Int. J. Fatigue 156, 106703 (2022).
This study was financially supported by Liaoning Provincial Research Foundation for Basic Research of China (Nos. 2023JH2), and the Fundamental Research Funds for the Central Universities (Nos. 04442023128), and the Key Laboratory of Mechanical Structure Optimization & Material Application Technology of Luzhou (Nos. SCHYZSB-2023-07), and National Nature Science Fund Project, Fracture toughness and mechanism of ductile iron material for large nuclear spent fuel storage and transportation container, Nos: 51174068, all support is gratefully acknowledged. There are no special ethical concerns arising from the use of animals or human subjects.
School of Intelligent Manufacturing and Automotive Engineering, Luzhou Vocational & Technical College, Luzhou, 646000, China
College of Computer Science and Engineering, Dalian Minzu University, Dalian, 116600, China
Hongcheng Zhang, Junxing Zhang & Hai Guo
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
You can also search for this author in PubMed Google Scholar
L.S. and C.H.Z. wrote the main manuscript text, X.J.Z. and H.G. collected literature and data, L.S. prepared Figs. 1, 2, 3, 4, 5, 6, 7 and 8 and C.H.Z. and H.G. prepared Figs. 9, 10, 11, 12, 13, 14, 15, 16 and 17. All authors reviewed the manuscript.
The authors declare no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
Song, L., Zhang, H., Zhang, J. et al. Prediction of heavy-section ductile iron fracture toughness based on machine learning. Sci Rep 14, 4681 (2024). https://doi.org/10.1038/s41598-024-55089-3
DOI: https://doi.org/10.1038/s41598-024-55089-3
Anyone you share the following link with will be able to read this content:
Sorry, a shareable link is not currently available for this article.
Provided by the Springer Nature SharedIt content-sharing initiative
Scientific Reports (Sci Rep) ISSN 2045-2322 (online)
compressor slide valve Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.