Nonlinear projection to latent structures method and its applications

被引:34
作者
Zhao, Shi Jian
Zhang, Jie [1 ]
Xu, Yong Mao
Xiong, Zhi Hua
机构
[1] Univ Newcastle Upon Tyne, Sch Chem Engn & Adv Mat, Newcastle Upon Tyne NE1 7RU, Tyne & Wear, England
[2] Tsing Hua Univ, Dept Automat, Beijing 100084, Peoples R China
关键词
D O I
10.1021/ie0512340
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
Projection to latent structures or partial least squares (PLS) is one of the most powerful linear regression techniques to deal with noisy and highly correlated data. To address the inherent nonlinearity present in the data from many industrial applications, a number of approaches have been proposed to extend it to the nonlinear case, and most of them rely on the nonlinear approximation capability of neural networks. However, these methods either merely introduce nonlinearities to the inner relationship model within the linear PLS framework or suffer from training a complicated network. In this paper, starting from an equivalent presentation of PLS, both nonlinear latent structures and nonlinear reconstruction are obtained straightforwardly through two consecutive steps. First, a radial basis function (RBF) network is utilized to extract the latent structures through linear algebra methods without the need of nonlinear optimization. This is followed by developing two feed-forward networks to reconstruct the original predictor variables and response variables. Extraction of multiple latent structures can be achieved in either sequential or parallel fashion. The proposed methodology thus possesses the capability of explicitly characterizing the nonlinear relationship between the latent structures and the original predictor variables while exhibiting fast convergence speed. This approach is appealing in diverse applications such as developing soft sensors and statistical process monitoring. It is assessed through both mathematical example and monitoring of a simulated batch polymerization reactor.
引用
收藏
页码:3843 / 3852
页数:10
相关论文
共 51 条
[1]   Asymptotic statistical theory of overtraining and cross-validation [J].
Amari, S ;
Murata, N ;
Muller, KR ;
Finke, M ;
Yang, HH .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1997, 8 (05) :985-996
[2]   PLS regression methods [J].
Höskuldsson, Agnar .
Journal of Chemometrics, 1988, 2 (03) :211-228
[3]  
[Anonymous], P IEEE INT C NEUR NE
[4]  
[Anonymous], P 1993 INT S INT CON
[5]   Non-linear projection to latent structures revisited: the quadratic PLS algorithm [J].
Baffi, G ;
Martin, EB ;
Morris, AJ .
COMPUTERS & CHEMICAL ENGINEERING, 1999, 23 (03) :395-411
[6]   Non-linear projection to latent structures revisited (the neural network PLS algorithm) [J].
Baffi, G ;
Martin, EB ;
Morris, AJ .
COMPUTERS & CHEMICAL ENGINEERING, 1999, 23 (09) :1293-1307
[7]   NEURAL NETWORKS AND PRINCIPAL COMPONENT ANALYSIS - LEARNING FROM EXAMPLES WITHOUT LOCAL MINIMA [J].
BALDI, P ;
HORNIK, K .
NEURAL NETWORKS, 1989, 2 (01) :53-58
[8]   LEARNING IN LINEAR NEURAL NETWORKS - A SURVEY [J].
BALDI, PF ;
HORNIK, K .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1995, 6 (04) :837-858
[9]   SOME THEOREMS ON QUADRATIC FORMS APPLIED IN THE STUDY OF ANALYSIS OF VARIANCE PROBLEMS .1. EFFECT OF INEQUALITY OF VARIANCE IN THE ONE-WAY CLASSIFICATION [J].
BOX, GEP .
ANNALS OF MATHEMATICAL STATISTICS, 1954, 25 (02) :290-302
[10]   An iterative pruning algorithm for feedforward neural networks [J].
Castellano, G ;
Fanelli, AM ;
Pelillo, M .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1997, 8 (03) :519-531