Development of Node-Decoupled Extended Kalman Network Diagnostic/Prognostic Reasoners
Download ReportTranscript Development of Node-Decoupled Extended Kalman Network Diagnostic/Prognostic Reasoners
Development of Node-Decoupled Extended Kalman Filter (NDEKF) Training Method to Design Neural Network Diagnostic/Prognostic Reasoners EE645 Final Project Kenichi Kaneshige Department of Electrical Engineering University of Hawaii at Manoa 2540 Dole St. Honolulu, HI 96822 Email: [email protected] Contents Motivation What is Kalman Filter? – Linear Kalman Filter – – – Kalman Filter and Neural Network Extended Kalman Filter (EKF) Node-Decoupled Extended Kalman Filter (NDEKF) Diagnosis/prognosis Simulation – – Result Detecting the fault condition – Node-decoupling NDEKF Algorithm Training the network – Simulation with Linear KF Create a situation Train the neural network Result Conclusion and possible future work Reference Motivation The detection of the system condition in a real time manner The node-decoupled extended kalman filter (NDEKF) algorithm Should work even when the input changes (robustness) The strength of the neural network What is Kalman Filter? Kalman filter = an optimal recursive data processing algorithm Used for stochastic estimation from noisy sensor measurements Predictor-Corrector type estimator Optimal because it incorporates all the provided information (measurements) regardless of their precision Linear Kalman Filter Used for linear system model Time update equation (predictor equation) – Responsible for projecting the current state the error covariance estimates – Obtain a priori estimates Measurement update equation (corrector equation) – Responsible for feedback Incorporate a new measurement into the a priori estimate – Obtain a posteriori estimates Linear Kalman Filter Algorithm Time Update (predict) Measurement Update (correct) (1) Project the state ahead xˆ k (1) Compute the Kalman gain Pk APk 1 AT Q Ax ˆ k 1 BU k (2) Update estimate with measurement Zk (2) Project the error covariance ahead K k Pk H T ( HPk H T R) 1 xˆ k xˆk k k ( zk Hxˆk ) (3) Update the error covariance k Pk ( I K k H ) P Initial estimates for xˆ k 1 and Pk 1 Simulation with Linear KF Linear Kalman Filter Performance 80 70 60 50 Position(feet) 40 30 20 10 0 Actual Altitude of the Aircraft Linear KF Estimate Sensor Data -10 -20 0 5 10 15 Time(sec) 20 25 30 The downside of LKF LKF seems to be working fine. What’s wrong? – Works only for the linear model of a dynamical system When the system is linear, we may extend Kalman filtering through a linearization procedure Kalman Filter and Neural Network Want better training methods in – – – – Training speed Mapping accuracy Generalization Overall performance The most promising training methods to satisfy above. – weight update procedures based upon second-order derivative information (Standard BP is based on first derivative) Popular second-order methods – Quasi-Newton – Levenburg-Marquardt – Conjugate gradient techniques However, these often converge to local optima because of the lack of a stochastic component in the weight update procedures Extended Kalman Filter (EKF) Second-order neural network training method During training, not only the weights, but also an error covariance matrix that encodes second-order information is also maintained Practical and effective alternative to the batch oriented Developed to enable the application of feedfoward and recurrent neural network (late 1980s) Shown to be substantially more effective than standard BP (epochs) Downside of EKF: computational complexity – Because of second-order information that correlates every pair of network weights Decoupled Extended Kalman Filter (DEKF) EKF: develops and maintains correlations between each pair of network weights DEKF: develops and maintains secondorder information only between weights that belong to mutually exclusive groups Family of DEKF: – Layer Decoupled EKF – Node Decoupled EKF – Fully Decoupled EKF Neural Network’s Behavior Process Equation: wk 1 wk k Measurement Equation: yk hk ( wk , uk , vk 1 ) k Process Noise: k ~ N [0, Q] E[klT ] k ,l Qk Measurement Noise: k ~ N [0, R] E[ k lT ] k ,l Rk wk Weight Parameter Vector yk Desired Response Vector hk Nonlinear Function uk Input Vector Node-decoupling Perform the Kalman recursion on a smaller part of the network at a time and continue until each part of the network is updated. Reduces the P() matrix to diagonal matrix State Variable Representation is the following xk 1 xk yk h( xk , uk ) k NDEKF Algorithm K ki Pki H ki Ak xˆ () Estimated value of x () xˆki 1 xki K ki k P() Approximate conditional error Pki1 Pki K ki H 'ik Pki Qk K () Kalman filter gain matrix N Ak [ Rk H 'ik Pki H ki ]1 i 1 covariance matrix of x () () Error between desired and actual outputs R() Measurement noise covariance matrix Q () Process noise covariance matrix H 'ik Weight update equation (take partial derivatives) (For details, please refer to the paper) NDEKF with Neural Network 1-20-50-1 MLFFNN is used. (1 input, 20 nodes in the first layer, 50 nodes in the second layer) Used bipolar activation function for the two hidden layers Linear activation function at the output node Training the Network Input Actual System Neural Network System in normal condition System in normal condition System in failed condition 1 System in failed condition 1 System in failed condition 2 System in failed condition 2 System in failed condition n System in failed condition n Simulation Assume the system or the plant has characteristics with following nonlinear equation y (k 1) 0.3 y (k ) 0.6 y (k 1) f [u (k )] f [u(k )] u 3 0.3u 2 0.4u Inputs are the following 2k For k=1:249 u (k ) sin( ) 250 For k=250:500 u (k ) 0.8 sin( 2k 2k ) 0.2 sin( ) 250 25 Why Neural Network? Input independency Robustness High accuracy for fault detection and identification Result for normal condition and failure condition 1 The Neural Network Output for Normal Condition 10 Amplitude Input Desired Output NN Output 5 0 -5 0 50 100 150 200 250 Time 300 350 400 450 500 The Neural Network Output for Failure Condition 1 10 Amplitude Input Desired Output NN Output 5 0 -5 0 50 100 150 200 250 Time 300 MSE = 4.7777 350 400 450 500 Result for failure condition 2 and 3 The Neural Network Output for Failure Condition 2 10 Amplitude Input Desired Output NN Output 5 0 -5 0 50 100 150 200 250 Time 300 350 400 450 500 The Neural Network Output for Failure Condition 3 10 Amplitude Input Desired Output NN Output 5 0 -5 0 50 100 150 200 250 Time 300 MSE=7.8946 350 400 450 500 The actual outputs of the system with its inputs. The Actual Output of the System 10 Amplitude Input Actual Output 5 0 -5 0 50 100 150 200 250 Time 300 350 400 450 500 Diagnosis of the system in actual condition Create an actual situation using one of the conditions (normal, failed1, failed2, failed3) Take MSE with the neural network of each of the conditions with the same input to the actual system. Take minimum of the MSE (MMSE), and it is the most probable condition of the actual system Result Here, the actual condition was tested with failed condition 2 Inputs 1 (Time 1 to 249) Inputs 2 (Time 250 to 500) Normal Condition 3.2703 1.3738 Failure Condition 1 5.4838 3.3171 Failure Condition 2 0.0123 0.1010 Failure Condition 3 5.4671 3.4935 From above, the MMSE shows the actual system is most probably be in failed condition 2 Conclusion and possible future work – The downside is there have to be a priori knowledge about the fault conditions – Work in frequency domain (FFT) – Implement with different algorithm and compare SVM, BP, Perceptron, etc… – Work with huge noise – Work with an actual model – OSA/CBM (Open Systems Architecture / Condition Based Maintenance) Using XML and let the real time report available on the internet References [1] Haykin, Simon Kalman Filtering and Neural Network John Wiley & Sons, Inc., New York 2001 [2] Murtuza, Syed; Chorian, Steven “Node Decoupled Extended Kalman Filter Based Learning Algorithm For Neural Networks”, IEEE International Symposium on Intelligent Control, August, 1994 [3] Maybeck, Peter “Stochastic models, estimation, and control; Vol. 1”, Academic Press 1979 [4] Narendra, K.S.; Parthasarathy, K.; “Identification and Control of Dynamical Systems Using Neural Networks”, IEEE Transactions on Neural Networks Volume: 1 Issue 1, Mar 1990 [5] Welch, Greg; Bishop Gary “An Introduction to the Kalman Filter”, Siggraph 2001, University of North Carolina at Chapel Hill [6] Ruchti, T.L.; Brown, R.H.; Garside, J.J.; “Kalman based artificial neural network training algorithms for nonlinear system identification” Intelligent Control, 1993., Proceedings of the 1993 IEEE International Symposium on , 25-27 Aug 1993 Page(s): 582 -587 [7] Marcus, Bengtsson, “Condition Based Maintenance on Rail Vehicles”, 2002 Technical Report [8] Wetzer, J.M.; Rutgers, W.R.; Verhaat, H.F.A. “Diagnostic- and Condition AssessmentTechniques for Condition Based Maintenance” 2000 Conference on Electrical Insulation and Dielectric Phenomena (cont’d) [9] Engel, Stephen; Gilmartin, Barbara; “Prognostics, The Real Issues Involved With Predicting Life Remaining” Aerospace Conference Proceedings, 2000 IEEE , Volume: 6 , 2000 Page(s): 457 -469 vol.6 [10] Hu, X.; Vian, J.; Choi, J.; Carlson, D.; Il, D.C.W.; “Propulsion vibration analysis using neural network inverse modeling” Neural Networks, 2002. IJCNN '02. Proceedings of the 2002 International Joint Conference on , Volume: 3 , 2002 Page(s): 2866 -2871 [11] Puskorius, G.V.; Feldkamp, L.A.; “Neurocontrol of nonlinear dynamical systems with Kalman filter trained recurrent networks” Neural Networks, IEEE Transactions on , Volume: 5 Issue: 2 , Mar 1994 Page(s): 279 -297 [12] Puskorius, G.V.; Feldkamp, L.A.; “Model reference adaptive control with recurrent networks trained by the dynamic DEKF algorithm” Neural Networks, 1992. IJCNN., International Joint Conference on , Volume: 2 , 7-11 Jun 1992 Page(s): 106 -113 vol.2 [13] Iiguni, Y.; Sakai, H.; Tokumaru, H.; “A real-time learning algorithm for a multilayered neural network based on the extended Kalman filter” Signal Processing, IEEE Transactions on , Volume: 40 Issue: 4 , Apr 1992 Page(s): 959 -966 [14] Jin, L.; Nikiforuk, P.N.; Gupta, M.M.; “Decoupled recursive estimation training and trainable degree of feedforward neural networks” Neural Networks, 1992. IJCNN., International Joint Conference on , Volume: 1 , 7-11 Jun 1992 Page(s): 894 -900 vol.1