ISIJ International
Online ISSN : 1347-5460
Print ISSN : 0915-1559
ISSN-L : 0915-1559
Instrumentation, Control and System Engineering
A Hybrid Modeling Method Based on Expert Control and Deep Neural Network for Temperature Prediction of Molten Steel in LF
Zi-cheng XinJiang-shan ZhangJin ZhengYu JinQing Liu
Author information
JOURNAL OPEN ACCESS FULL-TEXT HTML

2022 Volume 62 Issue 3 Pages 532-541

Details
Abstract

The temperature control of molten steel in ladle furnace (LF) has a critical impact on steelmaking production. In this work, production data were collected from a steelmaking plant and a hybrid model based on expert control and deep neural network (DNN) was established to predict the molten steel temperature in LF. In order to obtain the optimal DNN model, the trial and error method was used to determine the hyperparameters. And the optimal architecture of DNN model corresponds to the hidden layers of 4, hidden layer neurons of 35, iterations of 3000, and learning rate of 0.2. Compared with the multiple linear regression model and the shallow neural network model, the DNN model exhibits stronger generalisation performance and higher accuracy. The coefficient of determination (R2), correlation coefficient (r), mean square error (MSE), and root-mean-square error (RMSE) of the optimal DNN model reached 0.897, 0.947, 2.924, 1.710, respectively. Meanwhile, in the error scope of temperature from −5 to 5°C, the hit ratio of the hybrid model acquired 99.4%. The results demonstrate that the proposed model is effective to predict temperature of molten steel in LF.

1. Introduction

With the development of industrial intelligence and the increasing pressure of environmental protection, the intelligence, greenization and low carbonization have been paid more attention to the transformation and upgrading of steelmaking plants. Ladle furnace (LF) has been widely used in steelmaking plants for its low equipment investment and outstanding refining performance. The functions of LF refining include deoxidation, desulfurization, inclusion removal, adjustment of temperature and composition of molten steel, and homogenization of molten steel temperature and composition. The other main function is to ensure the smooth running of continuous casting process by controlling the temperature of molten steel. Therefore, it is of great significance to conduct accurate prediction of molten steel temperature, which is of benefit to the intelligent control of LF refining process.1,2)

Many researchers have studied the temperature prediction of molten steel in LF. There are currently three strategies to calculate or predict this temperature in the literature. One is to develop mechanism models and then obtain the correlation between temperature and the influencing parameters by using energy conservation equation, heat transfer equation, and mass conservation equation. Wu et al.3) established the mechanistic model of the heating rate of molten steel by using the law of energy conservation. Nath et al.4) developed a LF on-line reckoner based on simplified physics, material and heat balance, and statistical analysis of plant data to predict temperature and composition of steel. However, due to the complexity of LF refining process, many key technological parameters cannot be obtained directly, such as heat transfer coefficient of ladle furnace shell, temperature drop coefficient of argon stirring, heat loss coefficient of slag surface, which makes it difficult for the mechanism model to achieve accurate prediction. The second strategy is based on the intelligent algorithm to establish the nonlinear correlation between temperature and the influencing parameters. Tian et al.5,6,7,8) predicted the temperature of molten steel in LF by using extreme learning machine (ELM), back propagation (BP) neural network, and modified adaptive boosting (AdaBoost) algorithm. Lü et al.9) predicted the temperature of molten steel in LF based on optimally pruned bagging combined with partial linear extreme learning machine (PLELM), indicating the prediction accuracy of the model is higher by comparing with genetic algorithm-back propagation (GA-BP) model, partial least squares-support vector machine (PLS-SVM) model, and AdaBoost. RT-ELM model. However, the intelligent model excessively relies on data, which makes the model sensitive to abnormal data and vulnerable to the influence of imbalance data. The third strategy is based on combining the mechanism model and the intelligent algorithm to establish the hybrid model. Tang et al.10) and Fu et al.11) established a gray box model to predict the end-point temperature of molten steel in LF based on the mechanistic model and the black box model, indicating the prediction accuracy of the model is higher, which provides guidance for the LF refining process. However, the BP, SVM and ELM algorithm belong to shallow neural networks (SNN), and their limitations are mainly reflected in that not only the expression ability of complex functions in finite samples is limited, but also the generalization ability is restricted to some extent.12)

With the rapid development of artificial intelligence technology, another widely used burgeoning algorithm is deep neural networks (DNN).13) Kwon et al.14) predicted the hot ductility of steels based on elemental composition and thermal history by using the DNN indicating superior to the previous studies. Myers et al.15) predicted the nucleation lag time based on the elemental composition and the temperature for iron and steelmaking slags by using DNN. The DNN algorithm highlights the importance of feature learning. Through multi-layer nonlinear feature transformation, the feature representation of the sample in the original space is transformed into a new feature space. Meanwhile, each hidden layer can learn part of the data features, which can make the network store more information. The DNN algorithm can solve more complex data mapping problem and has stronger expression ability of complex function and generalization ability, which is of great benefit to the accurate prediction of molten steel temperature in LF refining process.

In this study, a hybrid model based on the expert control and the DNN algorithm was established to predict the temperature of molten steel in LF. The original production data collected from a steelmaking plant in China were preprocessed to delete outliers, and then the correlation analysis between the input and output variables was carried out. Moreover, the hybrid model was evaluated and compared with other models by using different evaluation indexes. It is expected to establish a more accurate and applicatory hybrid model to predict the temperature of molten steel in LF.

2. Analysis of LF Refining Process

2.1. Description of LF Refining Process

The steelmaking process includes BOF→LF→CC. The specific flow chart of LF refining process is shown in Fig. 1. When the ladle reached the LF station, the starting temperature of molten steel was measured, and a high-basicity reducing slag was obtained by adding slag making materials and deoxidizer. During the LF refining process, slag making materials and alloy are added to adjust the final composition of molten steel. When the composition of molten steel is qualified, the calcium packed wire is fed to realize the inclusion modification. Finally, the end temperature of molten steel is measured before the ladle leaves the LF station.

Fig. 1.

The specific flow chart of LF refining process.16) (Online version in color.)

2.2. Analysis Conservation of Energy

According to the law of conservation of energy, the energy can be divided into heat income and heat output in LF refining. The heat income comes from the heat of electric arc, the chemical reaction heat of additions. The heat output includes the heat exchanges between ladle furnace and surroundings, the heat storage of ladle shell, the heat loss from the slag surface, the heat loss by argon stirring, and the heat required for heating molten steel, as shown in Fig. 2. The Qarc is the heat of electric arc; the Qaddition is the chemical reaction heat of additions; the Qshell is the heat exchanges between ladle furnace and surroundings; the Qstorage is the heat storage of ladle; the Qsurface is the heat loss from the slag surface; the Qargon is the heat loss by argon stirring; the Qsteel is the heat change of molten steel. The LF can be seen as a system, and the heat income and the heat output of LF refining can reach a balance, as shown in Eq. (1).   

Q arc + Q addition = Q surface + Q argon + Q shell + Q storage + Q steel (1)
Fig. 2.

The heat budget of LF refining. (Online version in color.)

According to the relationship between input and output of energy, the heat change of molten steel can be expressed as Eq. (2).   

Q steel = Q arc + Q addition - Q surface - Q argon - Q shell - Q storage (2)

2.3. Analysis of Main Factors

Based on the actual operating conditions and the analysis of energy conservation of LF refining process, the main factors affecting the temperature prediction of molten steel were put forward. As shown in Table 1, the main factors include the addition amount of alloy, the addition amount of slag making materials, the turnover cycle of ladle, the weight of molten steel, the starting temperature of molten steel, the refining time, the heating duration time, and the argon consumption.

Table 1. Main factors affecting the temperature prediction of molten steel.
VariablesDescription of variablesUnits
X1Addition amount of alloykg
X2Addition amount of slag making materialskg
X3Turnover cycle of ladlemin
X4Weight of molten steelkg
X5Starting temperature of molten steel°C
X6Refining timemin
X7Heating duration times
X8Argon consumptionNL

The quantity of heat can be classified to the heat incomes and the heat outgoings. Based on the actual production, the relationship between the quantity of heat and the main factors affecting the temperature prediction of molten steel is analyzed as follows.

  • ▶  The heat incomes
    • ●  Heat of electric arc (Qarc): In the heating phase the voltage and the current are set to 200 V and 35 kA, respectively. The Qarc is mainly related to electric-arc heating duration.
    • ●  Chemical reaction heat of additions (Qaddition): The weight and composition of molten steel are different in different heats, which lead to the different addition of alloy and slag making materials. Based on the temperature effect coefficients of additions, the Qaddition can be calculated according to the amount of alloy and slag making materials.

  • ▶  The heat outgoings
    • ●  Heat loss from the slag surface (Qsurface): The Qsurface is mainly influenced by the thickness of slag layer. The thin slag usually promotes the heat loss from the slag surface.17) Actually, the casting residue will be recycled to the ladle furnace in the refining process, and the slag thickness is larger. Therefore, the effect of different slag thickness on heat loss is ignored. The Qsurface is mainly reflected by the refining time.
    • ●  Heat loss by argon stirring (Qargon): The bottom argon blowing will induce the exposure of the melt and result in heat loss of molten steel, and the argon consumption is proportional to the temperature drop of molten steel. Therefore, the Qargon can be reflected by the argon consumption.
    • ●  Heat storage of ladle (Qstorage): The ladle itself will absorb and store the heat of the molten steel in the refining process, and the Qstorage is mainly influenced by the thermal status of ladle. The thermal status of ladle is reflected by the turnover cycle of ladle in steelmaking process.
    • ●  Heat exchanges between ladle furnace and surroundings (Qshell): The heat is inevitably lost to the surroundings from the ladle furnace due to the temperature difference. The Qshell will increase with time, which can be reflected by the refining time.

Here, the starting temperature is the temperature measured for the first time after the ladle enters the LF station. The end temperature is the temperature measured for the last time before the ladle leaves the LF station.

3. Temperature Calculation of Molten Steel Using Hybrid Model

3.1. Modelling in Hybrid Model

A hybrid model based on metallurgical mechanism, expert control and DNN is established by analyzing the LF refining process of against the background of 150 t LF of a steelmaking plant in China. In this study, 2000 groups of production data were divided into two categories, of which the 1500 groups were used to train the hybrid training model, and the left 500 groups were used to test the hybrid model when the error between the predicted value (Tpredicted) of training model and the T1 reached the minimum. The T1 was the end temperature (Tmeasured) of molten steel minus the temperature change (ΔTaddition) caused by addition of alloy and slag making materials. The flowchart of construction for hybrid model is illustrated in Fig. 3.

Fig. 3.

A flowchart of construction for hybrid model. (Online version in color.)

3.1.1. Introduction of Expert Control

The concept of expert control proposed by K. J. Astrom, is to combine the theory and method of expert system with the control theory, and realize the control of the system by imitating the experience of experts under the unknown environment.18,19) Through field investigation, it can be seen that there are many kinds of slag making materials and alloy, and there are complex physical and chemical reactions in LF refining process. Meanwhile, the temperature effect coefficients of various additions were evaluated only by the experience of operators, which led to a large calculation error. Thus, a more effective control, that could give a reasonable quantity of heat generated by slag making materials and alloy when the end temperature calculated, was needed. The idea of expert control can solve the above problems well. In this study, the temperature effect coefficients of additions to molten steel, which was used to calculate the heat generated by slag-making material and alloy, was obtained by using the statistical analysis as shown in Table 2. And then, the temperature change of molten steel of various additions to molten steel was then calculated by using the idea of expert control, which combines the strengths of statistical analysis (obtaining the temperature effect coefficients), metallurgical mechanism (thermal equilibrium) and the experience of operators. The temperature change of molten steel owing to additions can be calculated using Eq. (3).   

Δ T addition = i G i q i (3)
Where, i designates a specific addition (alloy or slag making materials); Gi is the weight of addition i (kg); qi is the temperature effect coefficient of i (°C/kg).

Table 2. Temperature effect coefficients of various additions to molten steel in 150 t LF.
AdditionTemperature effect coefficient of addition to molten steel×10−2 (°C/kg)AdditionTemperature effect coefficient of addition to molten steel×10−2 (°C/kg)
C−4.30FeSi+1.10
HCFeMn−1.30Slag making materials−2.00
LCFeMn−1.15FeNb−0.32
CaSi−1.10FeTi−0.38

3.1.2. Introduction of DNN

Hinton et al.20) proposed the concept of deep learning (DL). The network structure of DNN is divided into input layer, hidden layer (multi-layer), output layer. Meanwhile, the layer and layer are directly fully connected.21) The DNN is used to predict temperature, the structure diagram of which is shown in Fig. 4. The input layer consists of six components, including turnover cycle of ladle, weight of molten steel, starting temperature of molten steel, refining time, heating duration time, and argon consumption. The output layer consists one neuron. The hyperparameter setting of DNN has an important effect on the prediction results, and the role of each hyperparameter will be described in detail below.

Fig. 4.

The structure diagram of DNN. (Online version in color.)

(1) Activation Function and Learning Rate

The activation function is a function that runs on the neurons of a neural network and is used to map the neural input to its output.22) Activation functions commonly used in DNN model include sigmoid (Sig), hyperbolic tangent (Tanh), and rectified linear unit (ReLU). For the DNN, the Sig and Tanh are prone to gradient disappearance, while the derivative of ReLU does not have this problem. Meanwhile, both Sig and Tanh need to calculate the exponential from the perspective of calculation, which is of high computational complexity. The ReLU only needs a threshold to get the activation value, and the rate of learning convergence is six times faster than the Sig or Tanh.23) However, ReLU units are fragile. When the gradient of input function is too large during training process, the neurons of ReLU no longer have the function to activate after the network parameters are updated, which results in the gradient will forever be zero.24) The leaky rectified linear unit (LReLU) don’t have this problem.14) Thus, the LReLU was selected as the activation function, as shown in Eq. (4). Learning rate is an important hyperparameter in DNN, which determines the change of weight generated in each training cycle.25,26) When the learning rate is large, the weight changes greatly, which easily leads to the instability of model. When the learning rate is small, the training time is increased and the convergence speed of model is slow.   

f lrelu ={ x 0.01x x>0 x0 (4)

(2) Number of Hidden Layers and Nodes

Both the number of hidden layers and nodes in hidden layer have great influence on the performance of DNN. The number of hidden layers mainly determines the complexity of network structure. The number of hidden layers of DNN commonly has 3 hidden layer (or >3 hidden layers).27) The number of hidden layer nodes mainly determines the learning ability and learning rate of DNN. When the number of hidden layer nodes is small, the learning ability and information processing ability of neural network are weak. When the number of hidden layer nodes is large, the learning rate of network will be slow, and it is easy to make the network fall into the local minimum value, which even leads to overfitting. The optimal number of hidden layer nodes and hidden layers was determined by using trial and error method.

(3) Optimization Algorithms

Gradient descent is one of the most popular optimization algorithms in DNN. The stochastic gradient descent (SGD) with momentum and L2 regularization was used to optimize the DNN, which not only solves the problems of slow convergence speed and easily getting into local optimum, but also prevents overfitting.28,29,30) The algorithm flow chart is shown in Table 3.

Table 3. Algorithm flow chart.
Algorithm 1: SGD with momentum and L2 regularization29,30)
Require: α: learning rate, β: momentum factor (Parameter defaults to 0.9), λ: L2 regularization factor (Parameter defaults to 1.0 e−4), f(θ): objective function with parameter θ
Require: θt=0(Initialize parameter vector)
t ← 0(Initialize timestep)
mt=0(Initialize moment vector)
ηt=0(Initialize schedule multiplier)
 while θt not converged do
  t ← t+1
  ▽ft(θt−1) ← θt−1(The gradient of θt−1 against loss)
  gt ← ∇ft(θt−1)+ λθt−1(Get gradients objective at timestep t)
  ηt ← Set schedule multiplier (t)
  mtβmt−1 + ηtαgt(Update biased moment estimate)
  θtθt−1mtηtλθt−1(Update parameters)
 end while stopping criterion is met
 return θt(Optimized parameters)

3.2. Analyzing of Database and Data

3.2.1. Data Cleansing

The production data of LF refining process were collected from a steelmaking plant in China. The original data were processed by using boxplot.31,32) The boxplot consists of five statistics, including lower quartile (Q1), median (Q2), upper quartile (Q3), Q1-1.5IQR, and Q3+ 1.5IQR. The interquartile range (IQR) is defined as Q3-Q1. In boxplot, the outlier data took values below Q1-1.5IQR or above Q3+ 1.5IQR, as shown in Fig. 5.

Fig. 5.

Outlier detection based on boxplot. (Online version in color.)

Boxplot is a widely used method to detect outliers in sample data. The advantage of boxplot is that it is not influenced by outliers, so it can describe the discrete distribution of data in a relatively stable way, and it is also conducive to data cleansing. In this study, the original data were processed using boxplot, and the outliers below Q1-1.5IQR and beyond Q3+ 1.5IQR were eliminated to realize the data cleaning. The outliers distribution of original data were shown after the original data was normalized, as shown in Fig. 6.

Fig. 6.

Outlier detection of boxplot based on original data. (Online version in color.)

There are 2000 groups of production data in total, as the data were obtained based on the boxplot and production operation rules. The data distribution of a single variable and the descriptive statistics for all input and output variables of the prediction models are showed in Fig. 7 and Table 4. The range of the turnover cycle of ladle and the refining time was 28–90 min and 18–55 min, respectively. The value of weight for molten steel ranged from 150000 to 158260 kg. The value of starting temperature ranged from 1518 to 1603°C. The range of the heating duration time was 155–1256 s. The range of the argon consumption was from 15000 to 48000 NL. The value of the T1 ranged from 1576 to 1612°C.

Fig. 7.

Scatter plot matrix visualization of X3, X4, X5, X6, X7, X8 and T1. (Online version in color.)

Table 4. Descriptive statistics of the variables.
VariablesMeanMinimumMaximumRange
Turnover cycle of ladle/min63289062
Weight of molten steel×105/kg1.52551.50001.58260.08260
Starting temperature×103/°C1.5551.5181.6030.0850
Refining time/min35185537
Heating duration time/s69315512561101
Argon consumption×104/NL3.01.54.83.3
T1×103/°C1.5931.5761.6120.0360

3.2.2. Correlation Analysis and Data Normalization

The correlation between two random variables was analyzed by using the Pearson correlation coefficient and the Student’s t-Test in the statistics analysis, as shown in Eqs. (5) and (6). In Fig. 8, the different impacts of input variables on the T1 of molten steel are shown in an order from strong to weak as: heating duration time, refining time, argon consumption, starting temperature of molten steel, turnover cycle of ladle and weight of molten steel. The Student’s t-Test was mainly used to test the significance of correlation coefficient. When the p-value is less than 0.01, it means that the correlation of variables is very significant. When the p-value of significance probability is less than 0.05, it means that the correlation of variables is significant. When the p-value is greater than 0.05, it means that the correlation of variables is not significant.33)   

r= i=1 n ( x i - x ¯ ) ( y i - y ¯ ) i=1 n ( x i - x ¯ ) 2 i=1 n ( y i - y ¯ ) 2 (5)
Where, x is the mean of variable x; y is the mean of variable y; xi is the ith value of variable x; yi is the ith value of variable y.   
t= r n-2 1- r 2 (6)
Where, r is the correlation coefficient; n is the sample size of variable; n−2 is the degree of freedom.
Fig. 8.

Correlation analysis results between input variables and output variables. (Online version in color.)

In this study, the p-value method was used to test the significance of the correlation between two variables. The results of correlation analysis are shown in Table 5. In Table 5, it can be seen that the p-value is less than 0.01 between T1 and starting temperature of molten steel, refining time, heating duration time, argon consumption, respectively; the p-value is less than 0.05 between T1 and turnover cycle of ladle; the p-value is greater than 0.05 between T1 and weight of molten steel. Taken together, these results indicated that the relationship between starting temperature of molten steel, refining time, heating duration time, argon consumption and T1 reach very significant correlation, and the relationship between turnover cycle of ladle and T1 reach significant correlation. However, there is no significant correlation between T1 and weight of molten steel. Actually, the weight of molten steel is non-negligible, which not only affects the addition of alloy and slag making materials but also affects the refining heating duration and the heat loss.6,7,8,34) Meanwhile, the initial energy of molten steel is determined by the weight of molten steel. Thus, the weight of molten steel was considered when the hybrid model was established.

Table 5. The calculation results of p-value between T1 and input variables.
X3X4X5X6X7X8
r−0.053−0.043−0.2180.3800.6450.292
p-value1.81×10−2*5.69×10−26.07×10−23**1.59×10−69**2.76×10−235**1.22×10−40**

Note: (*) p < 0.05, (**) p < 0.01. The sample size is 2000.

In order to improve the accuracy and accelerate the convergence of model, the data processed by using the boxplot and production operation rules were normalized. The mathematical formula of normalization processing is shown in Eq. (7).   

x * = x-min max-min (7)
Where, max is the maximum value of the data; min is the minimum value of the data.

3.3. Model Evaluation

The performance of hybrid model was evaluated according to different statistical evaluation indexes, including the coefficient of determination (R2), mean square error (MSE), and root-mean-square error (RMSE). The computational formulas of the above indexes are displayed in Eqs. (8), (9), (10):   

R 2 = i=1 N p ( y i act - y m ¯ ) 2 - i=1 N p ( y i pre - y i act ) 2 i=1 N p ( y i act - y m ¯ ) 2 (8)
  
MSE= i=1 N P ( y i pre - y i act ) 2 / N P (9)
  
RMSE= i=1 N P ( y i pre - y i act ) 2 / N P (10)
Where, Np is the number of the total data set; y is the value of the temperature; y m ¯ is the average value of the temperature; the “act” means the actual values; the “pre” means the predicted values.

4. Results and Discussion

4.1. Hyperparameter Optimization of DNN

In order to obtain the optimal DNN model, the trial and error method was used to determine the hyperparameters.35) To construct the optimal architectures of DNN model, the models were trained with varying the number of hidden layers. Considering the problem of training time, the initial hyperparameters of models were fixed as follows: the number of hidden layer neurons (20), the number of iterations (500) and learning rate (0.3). It can be seen from Fig. 9(a) that with the increase of the number of hidden layers, the MSE first decreased and then increased. When the number of hidden layers is 4, the MSE reaches the minimum. Hence, architectures consisting of 4 hidden layers were preferred for the DNN model. Figure 9(b) shows the effect of the number of hidden layer neurons on MSE at fixed hidden layer (4), iterations (500), and learning rate (0.3). It is evident from Fig. 9(b) that with the increase of the number of hidden layer neurons, the MSE decreased. When the number of hidden layer neurons exceeds 35, the MSE has little change. Then, at fixed hidden layer (4), hidden layer neurons (35), and learning rate (0.3), the MSE decreased with increasing iterations. When the iterations exceed 3000, the MSE has little change, as shown in Fig. 9(c). Figure 9(d) shows the variation of MES with learning rates at fixed hidden layer (4), hidden layer neurons (35), and iterations (3000). In Fig. 9(d), with the increase of the learning rate, the MSE first slightly decreased and then increased. The minimum MSE was obtained at learning rate of 0.2. To sum up, the optimal architecture of DNN model corresponds to the hidden layers of 4, hidden layer neurons of 35, iterations of 3000, and learning rate of 0.2.

Fig. 9.

MSE with variation of (a) the number of hidden layers, (b) the number of hidden layer neurons, (c) iterations, and (d) learning rate for DNN models. Note: The arrows indicate the optimal hyperparameters. (Online version in color.)

4.2. Comparison of the DNN Model with Other Models

Based on the Pearson correlation analysis, the relationship tends to be linear between the T1 and a part of input parameters, such as turnover cycle of ladle, starting temperature of molten steel, refining time, heating duration time, argon consumption respectively. However, there is a nonlinear feature between the weight of molten steel and the T1, which indicated that there are complex relationships (hybrid linear and nonlinear relationship) between input variables and output variables. Meanwhile, in the actual production, the weighing system has an error of more than 10 kg or even 100 kg; for example, the hopper weighing error is ±10 kg and the crown block weighing error is ±600 kg. In addition, the production datasets inevitably even contain fault data, abnormal data, or missing data, which led to the potential noisy data in the production dataset even after preprocessing. These are the two most important features of the data for modelling.

In order to verify the prediction accuracy and the generalization performance of the DNN model, the optimal DNN model were compared with multiple linear regression (MLR) model, back propagation (BP) neural network model, and regularized extreme learning machine (RELM) model. The performance evaluation results and prediction results of various models are shown in Tables 6, 7 and Fig. 10. In the MLR model, the R2 and the RMSE are 0.554, 3.556, respectively. In the shallow neural network (SNN) model, the R2 and the RMSE of the BP model and the RELM model are 0.668, 2.995 and 0.725, 2.729, respectively. These results show that the SNN model is better than the MLR model by comparing the R2 and the RMSE. The reason why the SNN model outperforms the MLR model is shown as follows: (1) The MLR is a relationship model between input variables (characteristic variables) and output variables. Since output variables are linear combinations of input variables, the MLR is always linear. The MLR does not fit well when the nonlinear data is processed. (2) All the data information of the SNN is stored in the neural unit, so the SNN has strong robustness. (3) The SNN model can adequately approximate any complex nonlinear relation. Therefore, the SNN model can get more accurate results when the input data are close to the training data of SNN model. (4) In the MLR models, the original features are directly used for modelling. However, in the SNN models, the original features of the input layer are processed by the hidden layer to obtain the new features, which is similar to the process of data pre-processing. Therefore, the prediction effect of the SNN model is better than that of the MLR model.36)

Table 6. Comparison of various models.
IndicesVarious models
MLRBPRELMDNN
R20.5540.6680.7250.897
r0.7440.8170.8520.947
MSE12.6428.9727.4462.924
RMSE3.5562.9952.7291.710

Table 7. Hit ratio of various models.
Error scopeVarious models
MLRBPRELMDNN
From −3 to 3°C63.2%70.4%76.8%91.4%
From −5 to 5°C85.4%90.4%93.4%99.4%
Fig. 10.

Comparison between T1 and Tpredicted obtained by (a) MLR, (b) BP, (c) RELM, and (d) DNN. The 45-degree diagonal dotted line is the identity where the T1 is equal to the Tpredicted. The inset is the distribution of the error scope. (Online version in color.)

The R2 and the RMSE of the DNN model are 0.897, 1.710, respectively. These results show that the DNN model is better than the SNN model by comparing the R2 and the RMSE. Since DNN not only has the advantages of SNN, but also has the advantages as follows: (1) Stronger learning ability. DNN takes the original data as input and gradually performs feature transformation through multiple hidden layers. Meanwhile, the rules of data are automatically extracted and stored in the network by learning, which can realize a more complex nonlinear mapping relationship.37,38) (2) Better generalization performance. As data structures become more complex and the amount of data increases, DNN has a stronger ability to process the high-dimensional data and express complex functions than SNN. Therefore, the DNN has a better generalization performance for complex problems.22) (3) Stronger anti-interference ability. In the case of large amount of data and complex data structure, the DNN will not have a great impact on the global training results when there are noisy data.

As shown in Fig. 10 and Table 7, the closer the scatter is to the 45-degree diagonal dotted line, the smaller the error between the T1 and Tpredicted will be. The overall scatter plot of DNN model is closer to the 45-degree diagonal dotted line than that of MRL model, BP model, and RELM model. In the error scope of temperature from −3 to 3°C, the hit ratio of MLR model, BP model, RELM model and DNN model acquired 63.2%, 70.4%, 76.8%, and 91.4%, respectively. In the error scope of temperature from −5 to 5°C, the hit ratio of MLR model, BP model, RELM model and DNN model acquired 85.4%, 90.4%, 93.4%, and 99.4%, respectively. The results show that the DNN model has higher accuracy and stronger generalization performance. Although hybrid model based on expert control and DNN has high prediction accuracy, there are still some shortcomings as follows: (1) Lack of theoretical support. Although there are a lot of optimization techniques for DNN models, there is still no relevant theory to support how to tune the hyperparameters, and the values of the hyperparameters are mainly determined by trial and error method. Therefore, the DNN models need to be further studied theoretically.39) (2) Computational efficiency problem. With the increase of data volume, the DNN model has higher prediction accuracy than the SNN model, but the DNN model has many hyperparameters, complex structure, large computing and other characteristics, which often leads to longer training time. Therefore, on the premise of ensuring the computational accuracy of DNN model, the improvement of computational efficiency should also be considered.40) (3) Engineering application problem. On the one hand, LF refining is a complicated physical and chemical process at high temperature, and there is a very complex nonlinear relationship between different variables. The complex process cannot be described completely and accurately by using the DNN model. Meanwhile, the expert control should be adjusted and optimized in real time to improve the hit ratio of the hybrid model, according to the change of the production condition, such as the change of the alloy and slag making materials supplier. On the other hand, in the actual production, the original production datasets inevitably contain fault data, abnormal data, and missing data, which will affect the calculation accuracy of the hybrid model. Therefore, the effective combination of hybrid model and big data mining technology should be considered to improve the data quality and then improve the calculation accuracy. After solving the above problems, the DNN model and the modeling strategy of the hybrid model can also be better applied in other metallurgy problems, which is of significance for the intelligent development of the iron and steel industry.

5. Conclusions

In this study, a hybrid modeling method based on expert control and deep neural network has been proposed for temperature prediction of molten steel in LF and the following conclusions can be drawn.

(1) Based on the boxplot and production operation rules, the original production data were preprocessed to delete outliers. In addition, the correlation analysis between the input and output variables was carried out. The results demonstrated that the different impacts of input variables on the T1 of molten steel are shown in an order from strong to weak as: the heating duration time (X7), the refining time (X6), the argon consumption (X8), the starting temperature of molten steel (X5), the turnover cycle of ladle (X3), and the weight of molten steel (X4).

(2) In order to obtain the optimal DNN model, the trial and error method was used to determine the hyperparameters. Meanwhile, the MSE of DNN model were considered when the DNN model was trained. The results demonstrated that the optimal architecture of DNN model corresponds to the hidden layers of 4, hidden layer neurons of 35, iterations of 3000, and learning rate of 0.2.

(3) The performance of the optimal DNN model was evaluated according to the statistical evaluation indexes and the distribution of error scope. The results demonstrated that the performance of the DNN model was better than that of the MLR model and SNN model; and the R2 and RMSE of the DNN model were 0.897, 1.710, respectively. Meanwhile, in the error scope from −5 to 5°C, the hit ratio of DNN model was 99.4%. To sum up, the hybrid model has higher accuracy and stronger generalisation performance to predict the temperature in LF refining process.

Acknowledgements

This work was financially supported by the National Natural Science Foundation of China (No. 51974023) and the funding of State Key Laboratory of Advanced Metallurgy, University of Science and Technology Beijing (No. 41621005).

References
 
© 2022 The Iron and Steel Institute of Japan.

This is an open access article under the terms of the Creative Commons Attribution-NonCommercial-NoDerivs license.
https://creativecommons.org/licenses/by-nc-nd/4.0/
feedback
Top