Download
artificial neural network chapter 5 back propagation network n.
Skip this Video
Loading SlideShow in 5 Seconds..
-Artificial Neural Network- Chapter 5 Back Propagation Network PowerPoint Presentation
Download Presentation
-Artificial Neural Network- Chapter 5 Back Propagation Network

-Artificial Neural Network- Chapter 5 Back Propagation Network

207 Views Download Presentation
Download Presentation

-Artificial Neural Network- Chapter 5 Back Propagation Network

- - - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

  1. -Artificial Neural Network-Chapter 5 Back Propagation Network Reference: 朝陽科大 李麗華 教授

  2. H1 Y1 X1 Y2 X2 H2 ‧ ‧ ‧ ‧ ‧ ‧ Hh ‧ ‧ ‧ ‧ ‧ ‧ Introduction (1) • BPN = Back Propagation Network • BPN is a layered feedforward supervised network. • BPN provides an effective means of allowing a computer to examine data patterns that may be incomplete or noisy. • Architecture: θ1 Wih Whj θ2 Yj Xn θh

  3. Y 1 Net Introduction (2) • Input layer: [X1,X2,….Xn]. • Hidden layer: can have more than one layer. • derive: net1, net2, …neth; transfer output H1, H2,…,Hh, to be used as the input to derive the result for output layer • Output layer: [Y1,…Yj]. • Weights: Wij. • Transfer function: Nonlinear Sigmoid function (*) The nodes in the hidden layers organize themselves in a way that different nodes learn to recognize different features of the total input space.

  4. Processing Steps (1) Briefly describe the processing Steps as follows. • Based on the problem domain, set up the network. • Randomly generate weights Wij. • Feed a training set, [X1,X2,….Xn], into BPN. • Compute the weighted sum and apply the transfer function on each node in each layer. Feeding the transferred data to the next layer until the output layer is reached. • The output pattern is compared to the desired output and an error is computed for each unit.

  5. Processing Steps (2) 5. Feedback the error back to each node in the hidden layer. • Each unit in hidden layer receives only a portion of total errors and these errors then feedback to the input layer. • Go to step 4 until the error is very small. • Repeat from step 3 again for another training set.

  6. x1 θ1 θj H1 net1 Wih : : Yj Whj : : Xi neth Hh : : θh Xn Computation Processes(1/10) • The detailed computation processes of BPN. • Set up the network according to the input nodes and the output nodes required. • Randomly assigned the weights. • Feed the training pattern (set) into the network and do the following computation.

  7. Computation Processes(2/10) 4. Compute from the Input layer to hidden layer for each node. 5. Compute from the hidden layer to output layer for each node.

  8. Computation Processes(3/10) 6. Calculate the total error & find the difference for correction δj=Yj(1-Yj)( Tj -Yj) δh=Hh(1- Hh) ΣjWhj δj 7. ΔWhj=ηδj Hh ΔΘj = -ηδj ΔWih=ηδh Xi ΔΘh= -ηδh 8. update weights Whj=Whj+ΔWhj,Wih=Wih+ΔWih, Θj= Θj + ΔΘj,Θh= Θh + ΔΘh 9. Repeat steps 4~8, until the error is very small. 10.Repeat steps 3~9, until all the training patterns are learned. Errors would be backward to revise

  9. W11 H1 X1 W13 W21 Θ1 Y1 W12 Θ3 H2 W23 X2 W22 Θ2 EX: Use BPN to solve XOR (1) • Use BPN to solve the XOR problem • Let W11=1, W21= -1, W12= -1, W22=1, W13=1, W23=1, Θ1=1, Θ2=1,Θ3=1, η=10

  10. EX: BPN Solve XOR (2) • ΔW12=ηδ1 X1 =(10)(-0.018)(-1)=0.18 • ΔW21=ηδ1 X2 =(10)(-0.018)(-1)=0.18 • ΔΘ1 =-ηδ1 = -(10)(-0.018)=0.18 • 以下為第一次修正後的權重值. 1.18 X1 0.754 0.82 0.82 1.915 0.754 X2 1.18

  11. BPN Discussion • Number of hidden nodes increase, the convergence will get slower. But the error can be minimized • The general concept of designing the number of hidden node uses: # of hidden nodes=(Input nodes + Output nodes)/2, or # of hidden nodes=(Input nodes * Output nodes)1/2 • Usually, 1~2 hidden layer is enough for learning a complex problem. Too many layers will cause the learning very slow. When the problem is hyper-dimension and very complex, then we add an extra layer. • Learning rate, η, usually set as [0.5, 1.0], but it depends on how fast and how detail the network shall learn.

  12. The Gradient Steepest Descent Method(SDM) (1) • The gradient steepest descent method • Recall: • We want the difference of computed output and expected output getting close to 0. • Therefore, we want to obtain so that we can update weights to improve the network results.

  13. The Gradient Steepest Descent Method(SDM) (2)

  14. The Gradient Steepest Descent Method(SDM) (3)

  15. The Gradient Steepest Descent Method(SDM) (4)

  16. The Gradient Steepest Descent Method(SDM) (5)

  17. The Gradient Steepest Descent Method(SDM) (6) • Learning computation

  18. The Gradient Steepest Descent Method(SDM) (7)

  19. 由美國職業高爾夫協會 122 位職業選手 1997 年的比賽記錄檔案中, 觀察其中的十二種高爾夫運動特有屬性來預測選手全年獎金之多寡 Case: (使用工具-Matlab)

  20. 匯入資料, 需事先將 Input_P.txt & Input_T.txt 放在 Work 目錄下 • load Input_P.txt; • load Input_T.txt; • p = Input_P'; % 調整 Input_P.txt 原始檔 • t = Input_T'; % 調整 Input_T.txt 原始檔

  21. 參考羅強華之 5-97 頁之 5.6.3 節之主成份分析, 目的為去除極端資料 • [pn, meanp, stdp, tn, meant, stdt] = prestd(p, t); % 將 p 和 t 標準化成 pn 和 tn • [ptrans, transMat] = prepca(pn, 0.05); % 只保留 95% 之資料來分析 • [R, C] = size(ptrans); % ptrans 為一 RxC 矩陣, 求 R = ? & C = ?

  22. 參考羅強華之 5-90 頁之 5.5.2 節之提早停止, 目的為改善網路廣義化 • set_trn = [1:4:C 3:4:C]; % 2/4 之訓練資料 • set_val = 2:4:C; % 1/4 之驗證資料 • set_tst = 4:4:C; % 1/4 之測試資料 • p_trn = ptrans(:,set_trn); % p_trn 之 Matrix • t_trn = tn(:,set_trn); % p_trn 之 Matrix • val.P = ptrans(:,set_val); % p_val 之 Structure, P 要大寫 • val.T = tn(:,set_val); % t_val 之 Structure, T 要大寫 • tst.P = ptrans(:,set_tst); % p_tst 之 Structure, P 要大寫 • tst.T = tn(:,set_tst); % t_tst 之 Structure, T 要大寫

  23. 建構倒傳遞網路 • net = newff(minmax(p_trn), [6 1], {'logsig', 'purelin'}, 'trainlm'); • [net, tr, y, e] = train(net, p_trn, t_trn, [], [], val, tst);

  24. 參考羅強華之 5-98 頁之 5.6.4 節之訓練後分析, 目的為分析性能結果 • an = sim(net, ptrans); % 把 ptrans 內的資料全部由網路模擬輸出 • a = poststd(an, meant, stdt); % 把本來已經標準化的資料全部還原 • for i = 1:1 • figure(i); • [m(i), b(i), r(i)] = postreg(a(i,:), t(i,:)); % 性能結果分析 • end