In the paper, we propose an SIRPP implementation on parallel processing with PVM, Parallel Virtual Machine.
In regression analysis, a lot of explanatory variables sometimes make data analysis harder. When some explanatory variables are useless for predicting the value of response variable, the explanatory variable selection is very useful. But, if all of the explanatory variables are related to the response variable, we will search linear combinations of explanatory variables with projection pursuit regression and ACE (Alternating Conditional Expectations). These two methods are assumed special models.
Sliced Inverse Regression (Li, 1991) is one of the approaches to reduce the number of explanatory variables in regression analysis. SIR does not get rid of some explanatory variables themselves but may reduce the dimension of a space of explanatory variables. It is based on the model (SIR model)
y=f(β
1x, β
2x, …, β
kx, ε),
where x is the vector of p explanatory variables, β
k(k=1, 2, …, K) are unknown row vectors, ε is independent of x, and f is an arbitrary unknown function on R[K+1].
Mizuta (1999) proposed an algorithm for SIR model with projection pursuit, named SIRPP (Sliced Inverse Regression with Projection Pursuit). SIRPP has excellent performance in finding {β
k}. However, the algorithm requires more computing power. It is a part of projection pursuit that the algorithm takes much time. In order to overcome the defect, we use PVM for SIRPP. We offer its effectiveness through numerical examples.
View full abstract