ISIJ International
Online ISSN : 1347-5460
Print ISSN : 0915-1559
ISSN-L : 0915-1559
Instrumentation, Control and System Engineering
A Mill Set-up Model Using a Multi-output Regression Tree for a Tandem Cold Mill Producing Stainless Steel
Hyun-Seok KangChi-Hyuck Jun
Author information
JOURNAL OPEN ACCESS FULL-TEXT HTML

2019 Volume 59 Issue 9 Pages 1582-1590

Details
Abstract

In a tandem cold mill for stainless steel, an optimum reduction rate is necessary for each stand. A conventional mill set-up uses a lookup-table to optimize the rolling schedule. However, to reflecting all the input conditions and manual interventions on a model is difficult. In this paper, we propose a mill set-up model that can efficiently predict the reduction rate for each stand by considering various input conditions. The proposed prediction model has a multi-output tree structure with a smaller time complexity for easy interpretation. The key contribution to the proposed algorithm is variable selection. According to the results of an analysis of the time-complexity, the proposed algorithm is less time consuming and is capable of learning datasets with a large number of variables more efficiently than the single-output CART (classification and regression trees). To evaluate the performance of the proposed algorithm, we applied it to the rolling reduction rate of a tandem cold mill in POSCO. The proposed algorithm achieves a similar level of R-squared in only 18% of the computing time required for an existing single-output CART algorithm.

1. Introduction

To increase productivity in the cold rolling of stainless steel, a tandem cold mill is commonly used. The general trend has been for the number of rolling mill stands to increase, with a few Chinese companies operating mills with more than five stands. Such configurations allow rolling mill companies to produce cold-rolled strips of stainless steel with higher productivity.

It is important to set up an optimum reduction rate for each stand in the tandem cold rolling of stainless steel. If the reduction rate of any one stand is excessive, it will cause surface defects, such as heat streaks.1) To prevent such defects in stainless steel cold-rolled plates, for which the surface quality of the material is important, the optimum reduction rate should be determined using a mill set-up model. Many studies2,3,4,5,6,7) have examined the setting up of the reduction rate using optimization techniques. There are several classifications for optimization techniques: 1) nonlinear programming methods,2) 2) GA algorithms,3) or 3) Simplex optimization.4,5,6) For the objective functions in optimization, previous studies have used uniform rolling load, motor power, shape, and flatness as the cost function.

One feature of tandem rolling is that each stand attains a high reduction rate because it is required to produce the target thickness with a limited number of rolling stands. The tandem rolling of stainless steels attains a relatively high rolling reduction rate, but large amounts of heat caused by friction and plastic deformation are generated. The high temperature causes the film to become thinner,8,9) which increases the probability of incurring heat-streak defects.10) Also, any roughness of the roll surface can give rise to defects.11) In other words, the optimization of the rolling reduction rate of a tandem cold mill for stainless steel should consider not only the rolling force balance between the stands, but also the temperature of the material, the film thickness, and the roughness of the rolls. All these cost functions are very complicated and difficult to implement in optimization algorithms. In actual rolling, the mill set-up model depends on the operators’ experience and their manual intervention. In this paper, we propose a reduction rate set-up model that learns operators’ experience from historical data.

The following sections in this paper are as follows. Section 2 describes the problems associated with a conventional mill set-up model and introduces the concept of a mill set-up learning model. Section 3 describes the various approaches to learning and explains the justification of the proposed learning model structure. Section 4 presents the extension of the single-output classification and regression trees (CART) to a multi-output CART and proposes an efficient algorithm to learn reduction rate of mill stand. Section 5 describes the datasets for learning the reduction rate of the actual rolling mill and analyzes the results. Finally, Section 6 presents the main conclusions.

2. Mill Set-up Models

2.1. Conventional Mill Set-up Model

Figure 1 is a schematic representation of the mill set-up model for a four-stand tandem mill. The target rolling reduction rate of each stand is based on the cold-rolling conditions (material type, target exit thickness, total reduction rate, etc.). Generally, the entry and exit thicknesses of the rolling mill are known variables. As such, the problem of setting up the reduction rate of each stand reduces to a problem of optimizing the target thickness between the rolling stands. For the cold rolling of stainless steel, we should consider the temperature of the material, the thickness of the oil film, and the surface roughness of the rolls in order to prevent surface defects such as heat streaks during the cold rolling. However, it is very difficult to reflect the above conditions in a mill set-up model for actual rolling.

Fig. 1.

Schematic of conventional mill set-up model. (Online version in color.)

In actual rolling, the algorithm for setting up the reduction rate is as follows; in this case, we consider a four-stand tandem cold mill.

First, the ratio of the reduction rates between stands is determined either empirically or experimentally, where ri is the reduction rate for the set-up at the i-th stand and ri′ is the table value (Eq. (1)).   

r 1 : r 2 : r 3 : r 4 = r 1 : r 2 : r 3 : r 4 (1)

The target reduction rate for each stand is defined by Eq. (2), where hi is the exit thickness at the i-th stand and Hi is the entry thickness at the i-th stand. Equation (3) shows the relationship between the entry and exit thicknesses at the neighboring stands.   

r i = H i - h i H i ,i=1,2,3,4 (2)
  
H i+1 = h i ,i=1,2,3 (3)

Then, we can determine the relationship between the entry and exit thicknesses (Eq. (4)). Finally, the thickness of the rolled steel can be expressed as the reduction ratio and entry thickness of the rolling mill (Eq. (5)).   

h i =( 1- r i ) H i ,i=1,2,3 (4)
  
h 4 =( 1+ r 1 ) ( 1+ r 2 ) ( 1+ r 3 ) ( 1+ r 4 ) H 1 (5)

There are four unknowns (ri, i = 1−4) in these equations, and we can establish three additional equations using the table values in Eq. (1).   

r 2 = r 2 r 1 r 1 (6)
  
r 3 = r 3 r 1 r 1 (7)
  
r 4 = r 4 r 1 r 1 (8)

By substituting Eqs. (6), (7), and (8) into Eq. (5), we can form a fourth-order polynomial equation of r1, given as Eq. (9), and this can be solved by applying a means of numerical analysis, such as the Newton-Raphson method. Then, the remaining reduction rate can be obtained from Eqs. (6), (7), and (8)   

h 4 =( 1+ r 1 ) ( 1+ r 2 r 1 r 1 ) ( 1+ r 3 r 1 r 1 ) ( 1+ r 4 r 1 r 1 ) H 1 (9)

In the above formula, the most important variable in the reduction rate set-up is the reduction ratio between the rolling stands. These values are usually determined from a lookup-table. However, this incurs several problems. First, when the number of input conditions increases, the lookup-table becomes excessively large. For example, if the input condition is divided into 10 steel types, 10 entry thickness groups, and 10 exit thickness groups, 1000 (10 × 10 × 10) lookups are generated, and it would be difficult to determine all these values experimentally. In addition, it is difficult for all the process conditions (such as annealing temperature and time) that affect the material properties to be reflected in the lookup-table. This is not only because the structure of the lookup-table has to be modified, but also because the size of the lookup table becomes proportionally greater due to the inclusion of the new conditions. In addition, the operators’ manual intervention is not reflected in the lookup table. Realistically, a lookup table can never be optimal for all actual rolling conditions, so we have to rely on manual intervention by the operators under certain conditions. However, since the lookup-table method is determined only by the rolling conditions, the degree of manual intervention cannot be reduced. This leads to variations among the operators.

2.2. A New Mill Set-up Model

To overcome the disadvantages of the existing mill set-up model, we propose a new alternative. Figure 2 is a schematic representation of the structure of this new model. A reduction rate prediction model replaces the lookup table. This prediction model can take previous process data as inputs as well as existing rolling data. Then, we predict the actual reduction rate by reflecting the manual intervention. It is possible to calculate the reduction ratio between stands based on the predicted reduction rate at each stand. This method can reflect the previous process data onto the set-up model, such that we expect that the model would be able to learn the reduction rate from the operator’s experience online. As a result, the degree of manual intervention would decrease.

Fig. 2.

Schematic of new mill set-up model. (Online version in color.)

3. Multi-output Prediction Models

3.1. Local and Global Approaches

The newly proposed reduction rate learning model can give rise to multi-output regression problems. There are two main approaches to multi-output regression.12) The first is the local method (Fig. 3(a)), which predicts the reduction rate of each stand independently. The second is the global method (Fig. 3(b)), which simultaneously predicts the reduction rates of all the stands. In the former case, a prediction model is constructed for each stand, and the output of each model has a scalar. In the latter case, a single model covers all the stands, and the output from the model has a vector in which the reduction ratio of each stand is a component.

Fig. 3.

Two types of multi-output regression. (a) Local method, (b) Global method.

The local method can utilize many existing algorithms, but the computation time can be longer than the global method. In addition, the local method cannot reflect the dependency between stands. However, the global method offers the advantage of being able to consider the dependency of the reduction rates between the stands at the same time.

3.2. Model Types

The different types of multi-output prediction models can be categorized as shown in Table 1. A linear regression model consists of a linear combination of inputs and model coefficients.13) Linear models are relatively easy to interpret, but even with a large number of input variables, this is not easy to achieve. In general, the performance is limited because the model does not reflect the nonlinearity. To reflect nonlinearity in the model, therefore, we can consider either a kernel method or a neural network model. In this case, however, the interpretation of the model is difficult because it requires the use of a nonlinear kernel function or activation function. When a black-boxed nonlinear model for the reduction rate prediction, it is difficult to interpret the model in actual implementation. One way of providing a white-boxed model for machine learning is to use a tree-structured model.14) The regression tree is a simple tree-like structure such as xi < c, where xi is a specific variable, and c is a split. This is as easy to interpret as the reduction ratio table. In this paper, we propose the multi-output, tree-structured model that would enable easy interpretation. Figure 4 shows an example of constructing a tree model in a global method. The data in the top node is divided into two groups, with each node minimizing the variance of the divided data. At the end node, the average value of the output of the divided data is sent to the output.

Table 1. Model characteristics of multi-output regression.
ModelCharacteristics
Linear regression modelsEasy interpretation
Low model flexibility and predictive accuracy
Support vector regressions (SVR)Easy interpretation, low model flexibility (Linear kernel function)
Hard interpretation, high model flexibility (Non-linear kernel function)
Regression treesVery easy interpretation
Fast for learning and prediction
Low on memory usage
Neural networksHard interpretation
High model flexibility
Hard to optimize the hyperparameters
Fig. 4.

Example of multi-output regression tree.

If we assume that the training data set D consists of N observations of input variable vector and output variable vector, D = {(x1, y1), …, (xN, yN)}. The l-th observation of the input variable vector is an m-dimensional vector x l =( x 1 l ,, x i l ,, x m l ) and the l-th observation of the output variable vector is a d-dimensional vector, y l =( y 1 l ,, y j l , y d l ) .

4. Proposed Algorithm for New Mill Set-up Model

4.1. Multi-output CART

CART is one of the most widely used algorithms.15) This method is used for single-output prediction and/or classification. The algorithm searches for a variable and a split to partition the data recursively, so as to reduce the impurity function as the tree grows. CART uses the Gini index for the impurity function in a classification problem, while the mean squared error (MSE) is used in the case of regression. The prediction of the reduction rate in our case is a regression problem, so we use MSE for the impurity function.

Suppose that the data set at node m (Qm) is partitioned into two sub-nodes (Qleft, Qright) by a split, as shown in Fig. 5. Also, Qm, Qleft, Qright have Nm, nleft, nright observations, respectively. The impurity function at node m before the split is calculated as:   

Impurity( Q m ) = j=1 d 1 N m l Q m ( y j l - y J ¯ ) 2 , (10)
where   
y J ¯ = 1 N m l Q m y j l (11)
Fig. 5.

Data partitioning at node m (Qm) using CART algorithm. (Online version in color.)

The decrease in the impurity function after applying the split s is expressed by   

ΔImpurity( s ) =Impurity( Q m ) - [ n left N m Impurity( Q left ) + n right N m Impurity( Q right ) ] (12)

Here, we need to find the best split s to maximize ΔImpurity(s).

The CART algorithm procedures are as follows. First, we sort for all continuous and ordinal x variables. Next, we exhaustively find the variable and split which minimize the impurity function. Finally, we divide the data into the derived variable and split. If the stopping criterion is not satisfied, the above steps repeat for the divided data. In the case of multi-output, all the steps are the same for single-output and only the impurity measure is different.16,17,18)

The time-complexity of the CART algorithm is O(mNlogN), where m is the number of variables and N is the number of observations.19) The learning time is proportional to the number of input variables. As described in Section 2.2, the input items of the newly proposed learning model include the data of the pre-cold rolling process. That is, if the number of input items is large, it will take a long time to learn the model when using the CART algorithm. To learn the model in the actual rolling process, a more efficient algorithm than that of the CART is needed.

4.2. Clustering-based Split-variable Selection

The key concept of the proposed algorithm is to use clustering when selecting a split variable. The most inefficient part of the CART algorithm is that all the input variables must be sorted before a split variable can be selected. The Quicksort algorithm has a time-complexity of O(NlogN) and will have total time-complexity of (mNlogN) if it is sorted for every variable. If we could find the variable in which the impurity function is minimum, not be necessary to sort all the variables. Therefore, a tree could be built more efficiently by sorting only a specific variable if we could find the variable giving the minimum impurity.

The CART partitions the data into two groups to minimize the specific impurity measure. This is similar to the process of a clustering algorithm. Among the clustering algorithms, one of the most efficient is the k-means algorithm. In other words, the k-means algorithm could provide an alternative method of building a tree-structured model.

The k-means algorithm clusters the data to minimize the within-cluster variance, where k is the number of clusters. If k = 2, the k-means algorithm minimizes the variance sum of two partitions of observations. In this case, upon applying the k-means algorithm to observations consisting of a pair of one arbitrary input variable and one output variable, the observations are clustered into two groups as a function of the input and output variables. The algorithm proposed in this paper attempts to attain the split-variable selection using this property.

To illustrate the principle of the variable selection algorithm based on the k-means algorithm, we used pairs of example data with one arbitrary input variable and one output variable (Data source: Boston housing data from the UCI repository of machine learning database,20) x: average number of rooms per dwelling, y: median value of owner-occupied homes in $1000). Figure 6 shows the scatter plot obtained using the example data. The figure shows that the relationship between the input and output is roughly linear. The goal is to divide the data into two clusters so that the output variance is minimized when using the k-means algorithm.

Fig. 6.

Scatter plot of x6y for Boston house price dataset. (Online version in color.)

By modifying the within-cluster distance measure, we can construct a modified k-means algorithm such as 1) and 2). If we sequentially apply steps 1) and 2) for each iteration of the k-means algorithm, we can construct a tree model that is likely to minimize the output within-cluster variance.

1) Distance measure defined along the y-axis

As shown in Fig. 7, the data is divided by the y-axis and the summation of the variance of y is minimized. This has the advantage of minimizing the MSE, which is an impurity measure of the CART, but it cannot be used to construct a model with a tree structure like xi < c.

Fig. 7.

Clustering result of x6y for Boston house price dataset. (Distance measure: distance along y-axis). (Online version in color.)

2) Distance measure defined along the x-axis

As shown in Fig. 8, the data is divided by the x-axis, and the summation of the variance of x is minimized. We can construct a general tree-structured model. However, it cannot guarantee that the variance of y is minimum.

Fig. 8.

Clustering result of x6y for Boston house price dataset. (Distance measure: distance along x-axis). (Online version in color.)

In the k-means algorithm, the center of each cluster is updated using the mean value. However, since the mean value is sensitive to outliers, it degrades the clustering performance. In particular, there is the possibility that outliers may exist in the process data in actual rolling, so there is a need for an algorithm that is less sensitive to outliers. According to Reference 21), the selection of the medoids closest to the mean, instead of the mean, when updating the reference point of each cluster, improves the clustering performance. The algorithm proposed in this paper adopts this concept.

Applying the algorithm mentioned above makes it easy to divide the data and obtain the impurity measure based on clustered data. Also, if we apply the algorithm to all the variables, we can determine the impurity measure for each variable and select the optimum variable with the lowest impurity.

4.3. Details of Proposed Algorithm

Figure 9(a) is a flow chart of the algorithm proposed in this paper. Once the learning data are obtained, a pair consisting of one variable and an output is defined as the input data for the clustering algorithm. Then, the clustering algorithm is applied to the pre-defined data to obtain the degree of impurity of the variable. After applying the above clustering algorithm to all the variables, we can identify which variable has the lowest level of impurity.

Fig. 9.

Flow chart of proposed algorithm. (a) Main flow chart of proposed algorithm for multi-output regression tree, (b) Proposed clustering algorithm for calculating degree of impurity of xi.

Figure 9(b) illustrates the clustering algorithm. First, the two centroids are initialized with the minimum and maximum values of the input variable and the output. Then, the distance between the two centroids and each observation along the output is calculated. All the observations can be assigned to the cluster for which the distance from the centroids is the smallest (as shown in Fig. 7). The means of the clustered data are calculated, and the resulting value is used to update the centroid. Once again, the distances between the new centroids and each observation along the input variable are calculated, after which each observation is assigned to the cluster in the same way as described above (as shown in Fig. 8). Afterwards, the centroids are updated with the medoids with the nearest means of each cluster. If there is no difference from the centroid in the previous step, the algorithm stops and the impurity measure is obtained.

After the selection of the variable, the subsequent procedure is the same as the existing CART algorithm. It sorts the selected variable and exhaustively searches for the split point at which the impurity is a minimum. Then, the data can be divided into two parts as a result of the obtained split. If the stopping criterion is not satisfied, the above steps are repeated for the divided data.

4.4. Time-complexity Analysis of the Proposed Algorithm

In this section, we analyze the time-complexity in detail in order to determine the time efficiency of the proposed algorithm. The proposed algorithm is divided into two steps. First, the variable selection part executes the clustering algorithm in much the same way as the k-means algorithm, as many times as for the input variable. In this algorithm, the time-complexity of the k-means algorithm is O(kNdniter), where k, N, niter, and d are the number of clusters, observations, iterations, and the dimension of the output, respectively.22) Since k is 2, d is 4 (4 stands), and niter is smaller than N, the total time-complexity of each clustering and variable selection algorithm is O(N), O(mN), respectively.

The second step involves searching for the split where the impurity is a minimum. Unlike the existing CART algorithm, it only sorts the selected variable, such that the time-complexity of this part is O(NlogN). Since the latter part is more time consuming, the total time-complexity of the proposed algorithm is O(NlogN). This algorithm is less affected by the number of variables than the existing CART algorithm. It can learn datasets with a large number of variables more efficiently.

5. Experiment and Results

5.1. Datasets

To evaluate the performance of the proposed algorithm, we applied it to the set-up of the rolling reduction rate of a tandem cold mill for stainless steel in POSCO. This process uses a four-stand mill and has been producing cold-rolled strips since 2009. Figure 10 shows the actual plant, and the approximate specifications of the mill are listed in Table 2. The entire process for producing cold-rolled strip is as follows. First, slabs are produced by steelmaking and continuous-casting processes, after which hot-rolled strips are produced through a hot-rolling process. Then, white-coils are produced through annealing and cold-rolled strip is produced through continuous cold rolling. Several processes must be completed before the cold-rolling process. The data for these processes are used as the input, and the actual reduction rates of each stand of the rolling mill are used as the outputs. The data for each process are listed in Table 3. The data were acquired in 2017, with 2200 coils being randomly selected.

Fig. 10.

Actual plant. (Online version in color.)

Table 2. Mill specification.
Specificationdata
Annual production500000 t/year
Material gradesStainless-steel hot and cold strip
Strip data, entry sectionWidth600 to 1350 mm
Thickness1.8 to 5.0 mm (Hot strip)
0.8 to 2.0 mm (Cold strip)
Coil weightMax. 40000 kg
Strip data, exit sectionWidth600 to 1350 mm
Thickness0.4 to 2.0 mm
Coil weightMax. 40000 kg

Table 3. Input, output data list.
input/outputprocessvariableitem
inputsteel makingx1wt% of C
x2wt% of Si
x3wt% of Mn
x4wt% of P
x5wt% of S
x6wt% of Cu
x7wt% of Al
x8wt% of Nb
x9wt% of Ni
x10wt% of Cr
x11wt% of Mo
x12wt% of Ti
x13wt% of Co
x14wt% of N
x15wt% of B
x16slab thickness
x17slab width
x18slab length
hot rollingx19hot-rolled strip thickness
x20hot-rolled strip width
x21hot-rolled strip length
x22slab temperature at furnace entry
x23temperature at preheating zone
x24temperature at heating zone
x25temperature at soaking zone
x26preheating time
x27heating time
x28soaking time
x29pitch time of finishing mill
x30roll unit order
x31number of pass in #1 roughing mill
x32number of pass in #2 roughing mill
x33entry temperature of finishing mill
x34exit temperature of finishing mill
x35coiling temperature
x36bar thickness
x37bar width
x38hot-rolled strip crown
annealingx39number of process
x40number of furnace
x41temperature at preheating zone
x42temperature at heating zone
x43temperature at soaking zone
x44reduction rate of skin pass mill
x45line speed
x46white coil width
cold rollingx47target thickness
x48total reduction rate
outputcold rollingy1#1 stand reduction rate
y2#2 stand reduction rate
y3#3 stand reduction rate
y4#4 stand reduction rate

5.2. Algorithm Comparison among Single-output Regression

This section compares the performances of several single-output regression models. By doing this, we can compare the characteristics of each model and identify the performance target of the proposed algorithm. The model types are divided into linear regression, regression tree, support vector regression (SVR), ensemble tree, Gaussian process regression (GPR), and neural network (NN). The forms of each model are shown in Table 4. The simulations of all these models were performed using Matlab® applications. (linear regression, regression tree, SVR, ensemble tree, and the GPR model were fitted by the Regression Leaner function, while the NN model was fitted by the Neural Network Fitting function). Since these are single-output regressions, the reduction rates of the four stands were determined independently. The learning performances of the different models were compared using R-squared (Table 5).

Table 4. Details of local methods for single-output regression.
Model groupModelModel options
Linear regressionMultiple regressionConstant and linear terms
Regression treeFine treeMax. leaf size is 4
Medium treeMax. leaf size is 12
Simple treeMax. leaf size is 36
SVRLinear SVRLinear kernel function
Ensemble treeBoosted treeLeast-squares boosting
Bagged treeBootstrap bagging
GPRMatern 5/2Matern 5/2 kernel function
ExponentialExponential kernel function
NNMulti-layer perceptron1 hidden layer : 10 neurons

Table 5. Simulation results for performance of local models.
Model groupModelR-squared [%]
y1y2y3y4
Linear regressionMultiple regression89.2481.4192.9494.39
Regression treeFine tree95.1183.6995.8297.36
Medium tree94.5884.8796.9296.73
Simple tree93.6985.0196.2995.92
SVRLinear SVR90.2579.6893.8194.04
Ensemble treeBoosted tree80.9776.1379.2679.88
Bagged tree95.9190.0997.4396.76
GPRMatern 5/295.6687.2296.5997.01
Exponential95.9487.5397.3596.83
NNMulti-layer perceptron93.2784.1793.4194.70

According to the simulation, the best-performing model was ensemble bagged tree. This was followed by GPR, regression tree, and the NN models. Overall, tree-structured models exhibited a relatively high level of performance. In general, a NN model exhibits a high learning performance. It is interesting the performance of the tree-structured model surpasses that of the NN model. This is because the learned data are obtained from the table. A level of approximately 92–95% can be judged as being the highest level of learning performance for all the stands, and can be set as the target level of performance for the proposed algorithm.

Ensemble models can learn and combine several weak learner models to achieve a higher level of performance.23,24) However, they are not recommended for implementation in an actual rolling mill model because the entire model is difficult to interpret and does not outperform the simple regression tree models by a particularly significant margin.

The NN model also exhibits a level of performance in excess of 90% R-squared. However, the model incurs several disadvantages that make it difficult to interpret, because it contains nonlinear activation functions, such as sigmoid and ReLu functions.25) In addition, the structure of this model includes many hyperparameters, making it difficult to find an optimal model.26,27) For example, the types of activation function and the number of hidden layers and neurons are not easy to optimize unless the actual learning performances are compared.

As a result of assessing the performance and interpretational aspects of the model, it is thought that the regression tree-structure models are suitable for learning the rolling reduction rate.

5.3. Results Obtained Using the Proposed Algorithm

In this section, we compare the results obtained with the proposed and existing CART algorithms, and examine the applicability to the actual rolling mill model. To fairly compare the learning performance, we implemented the proposed and CART algorithms in Python®. Learning was executed 100 times for each algorithm. In each case, we apportioned the data into training and test sets, with and 80-20 split. As the hyperparmeters of the model, the maximum depth of the tree is 1000 and the minimum number of samples required to be at a node is 5% of the number of the training data set. The average R-squared value and the learning time were obtained. The most important advantage of the proposed algorithm is its high time efficiency. Therefore, the performance was evaluated based on the R-squared and learning time values.

The proposed algorithm is applicable to both single- and multi-output regression tree. The performances of each model are listed in Table 6. The single-output regression tree learns the reduction rate of each stand, such that it has four sub-models. The learning time of this model is, on average, more than four times greater than that of a multi-output regression tree. Also, it is shown that the multi-output tree model learned by the proposed algorithm requires only about 18% of the learning time relative to the existing single-output CART algorithm. Therefore, it can be concluded that the proposed algorithm is better than CART in terms of time efficiency. In addition, in terms of R-squared, the multi-output regression tree is slightly inferior to the single-output regression model, but not by a significant margin. The proposed tree-structure model and algorithm can achieve a similar R-squared level of performance to the existing single-output CART algorithm.

Table 6. Performance of tree-type models.
Tree structureAlgorithmR-squared [%]Time [s]
y1y2y3y4All
Single-outputCART93.5685.4896.1496.2692.861978
Proposed93.5285.395.7396.0492.65362
Multi-output (Proposed)Multi-variate CART92.9385.3895.7695.7892.46468
Proposed92.8785.1595.3695.7292.2384

Figure 11 shows scatter diagrams of each stand’s actual and predicted reduction rate from the proposed model with a test dataset. As the scatter approaches the linear, the R-squared value increases. As such, we can see that the learning performance for the second stand is lower than for the other stands. This is similar to the results obtained with single-output regression and is due to the high rate of manual intervention in the second stand. In general, to prevent heat-streak defects, operators usually adjust the rolling reduction rate of the second stand. This reflects the actual operation of the rolling mill. Figure 12 is a time series chart of the reduction rate of the first stand. We can see that the difference between the actual and predicted values is negligible. However, there are 3 points (coil number are 14, 30 and 45) which have more than 3.5% difference between the actual and prediction values. We presume that these points are affected by manual interventions. Although the difference is excessively large, the reduction ratio among stands is calculated and applied to the set-up model.

Fig. 11.

Scatter plots of actual/predicted reduction rate for each stand (Dotted lines: actual reduction rate = predicted reduction rate). (Online version in color.)

Fig. 12.

Time series chart for actual, predicted reduction rate of #1 stand. (Online version in color.)

6. Conclusions

In this paper, we proposed a mill set-up model that efficiently predicts the reduction rate of a tandem cold mill for stainless steel. This model can be implemented using the actual reduction rate of each stand with the data obtained from cold rolling and the previous processes. Given that this is not a conventional lookup-table model, it can set up the reduction rate by incorporating operators’ experience online. Moreover, it is possible to reflect changes in the major engineering process variables in the set-up model. As such, it should be possible to reduce the variance and the degree of manual intervention by operators.

The proposed model is a multi-output regression tree. Therefore, it offers the advantage of building only a single model for the multiple-output variables and reflects the dependency of the reduction rates among stands. The existing single-output CART searches for a variable and a split to partition the data recursively, such that it is not time efficient. To overcome this issue, we proposed a more time-efficient split-variable selection algorithm based on clustering. This allows us to sort only one variable and search for a split that minimizes the impurity measure. According to the analysis of time-complexity, the proposed algorithm is less time consuming.

To evaluate the performance of the proposed algorithm, we applied it to the set-up problem of the rolling reduction rate of a tandem cold mill for producing stainless steel in POSCO. According to the simulation results, the proposed algorithm exhibits a similar level of R-squared performance and a higher time efficiency compared with the existing CART algorithm.

Acknowledgement

This work was supported by a grant from the National Research Foundation of Korea (Project Number: 2017R1A2B4005450). We would like to thank Stainless Steel Rolling Dept. in POSCO for providing with rolling data.

References
 
© 2019 by The Iron and Steel Institute of Japan
feedback
Top