top

  Info

  • Utilizzare la checkbox di selezione a fianco di ciascun documento per attivare le funzionalità di stampa, invio email, download nei formati disponibili del (i) record.

  Info

  • Utilizzare questo link per rimuovere la selezione effettuata.
A statistical approach to neural networks for pattern recognition [[electronic resource] /] / Robert A. Dunne
A statistical approach to neural networks for pattern recognition [[electronic resource] /] / Robert A. Dunne
Autore Dunne Robert A
Pubbl/distr/stampa Hoboken, N.J. ; ; Chichester, : Wiley, c2007
Descrizione fisica 1 online resource (289 p.)
Disciplina 006.32
Collana Wiley series in computational statistics
Soggetto topico Perceptrons
Neural networks (Computer science)
Soggetto genere / forma Electronic books.
ISBN 1-280-93517-0
9786610935178
0-470-14815-2
0-470-14814-4
Formato Materiale a stampa
Livello bibliografico Monografia
Lingua di pubblicazione eng
Nota di contenuto A Statistical Approach to Neural Networks for Pattern Recognition; Contents; Notation and Code Examples; Preface; Acknowledgments; 1 Introduction; 1.1 The perceptron; 2 The Multi-Layer Perceptron Model; 2.1 The multi-layer perceptron (MLP); 2.2 The first and second derivatives; 2.3 Additional hidden layers; 2.4 Classifiers; 2.5 Complements and exercises; 3 Linear Discriminant Analysis; 3.1 An alternative method; 3.2 Example; 3.3 Flexible and penalized LDA; 3.4 Relationship of MLP models to LDA; 3.5 Linear classifiers; 3.6 Complements and exercises; 4 Activation and Penalty Functions
4.1 Introduction4.2 Interpreting outputs as probabilities; 4.3 The fiuniversal approximatorfl and consistency; 4.4 Variance and bias; 4.5 Binary variables and logistic regression; 4.6 MLP models and cross-entropy; 4.7 A derivation of the softmax activation function; 4.8 The finaturalfl pairing and A,; 4.9 A comparison of least squares and cross-entropy; 4.10 Conclusion; 4.11 Complements and exercises; 5 Model Fitting and Evaluation; 5.1 Introduction; 5.2 Error rate estimation; 5.3 Model selection for MLP models; 5.4 Penalized training; 5.5 Complements and exercises; 6 The Task-based MLP
6.1 Introduction6.2 The task-based MLP; 6.3 Pruning algorithms; 6.4 Interpreting and evaluating task-based MLP models; 6.5 Evaluating the models; 6.6 Conclusion; 6.7 Complements and exercises; 7 Incorporating Spatial Information into an MLP Classifier; 7.1 Allocation and neighbor information; 7.2 Markov random fields; 7.3 Hopfield networks; 7.4 MLP neighbor models; 7.5 Sequential updating; 7.6 Example - MartinTMs farm; 7.7 Conclusion; 7.8 Complements and exercises; 8 Influence Curves for the Multi-layer Perceptron Classifier; 8.1 Introduction; 8.2 Estimators; 8.3 Influence curves
8.4 M-estimators8.5 The MLP; 8.6 Influence curves for pc; 8.7 Summary and Conclusion; 9 The Sensitivity Curves of the MLP Classifier; 9.1 Introduction; 9.2 The sensitivity curve; 9.3 Some experiments; 9.4 Discussion; 9.5 Conclusion; 10 A Robust Fitting Procedure for MLP Models; 10.1 Introduction; 10.2 The effect of a hidden layer; 10.3 Comparison of MLP with robust logistic regression; 10.4 A robust MLP model; 10.5 Diagnostics; 10.6 Conclusion; 10.7 Complements and exercises; 11 Smoothed Weights; 11.1 Introduction; 11.2 MLP models; 11.3 Examples; 11.4 Conclusion
11.5 Cornplernents and exercises12 Translation Invariance; 12.1 Introduction; 12.2 Example 1; 12.3 Example 2; 12.4 Example 3; 12.5 Conclusion; 13 Fixed-slope Training; 13.1 Introduction; 13.2 Strategies; 13.3 Fixing γ or O; 13.4 Example 1; 13.5 Example 2; 13.6 Discussion; Bibliography; Appendix A: Function Minimization; A.l Introduction; A.2 Back-propagation; A.3 Newton-Raphson; A.4 The method of scoring; A.5 Quasi-Newton; A.6 Conjugate gradients; A.7 Scaled conjugate gradients; A.8 Variants on vanilla fiback-propagationfl; A.9 Line search; A.10 The simplex algorithm; A.11 Implementation
A.12 Examples
Record Nr. UNINA-9910143416103321
Dunne Robert A  
Hoboken, N.J. ; ; Chichester, : Wiley, c2007
Materiale a stampa
Lo trovi qui: Univ. Federico II
Opac: Controlla la disponibilità qui
A statistical approach to neural networks for pattern recognition [[electronic resource] /] / Robert A. Dunne
A statistical approach to neural networks for pattern recognition [[electronic resource] /] / Robert A. Dunne
Autore Dunne Robert A
Pubbl/distr/stampa Hoboken, N.J. ; ; Chichester, : Wiley, c2007
Descrizione fisica 1 online resource (289 p.)
Disciplina 006.32
Collana Wiley series in computational statistics
Soggetto topico Perceptrons
Neural networks (Computer science)
ISBN 1-280-93517-0
9786610935178
0-470-14815-2
0-470-14814-4
Formato Materiale a stampa
Livello bibliografico Monografia
Lingua di pubblicazione eng
Nota di contenuto A Statistical Approach to Neural Networks for Pattern Recognition; Contents; Notation and Code Examples; Preface; Acknowledgments; 1 Introduction; 1.1 The perceptron; 2 The Multi-Layer Perceptron Model; 2.1 The multi-layer perceptron (MLP); 2.2 The first and second derivatives; 2.3 Additional hidden layers; 2.4 Classifiers; 2.5 Complements and exercises; 3 Linear Discriminant Analysis; 3.1 An alternative method; 3.2 Example; 3.3 Flexible and penalized LDA; 3.4 Relationship of MLP models to LDA; 3.5 Linear classifiers; 3.6 Complements and exercises; 4 Activation and Penalty Functions
4.1 Introduction4.2 Interpreting outputs as probabilities; 4.3 The fiuniversal approximatorfl and consistency; 4.4 Variance and bias; 4.5 Binary variables and logistic regression; 4.6 MLP models and cross-entropy; 4.7 A derivation of the softmax activation function; 4.8 The finaturalfl pairing and A,; 4.9 A comparison of least squares and cross-entropy; 4.10 Conclusion; 4.11 Complements and exercises; 5 Model Fitting and Evaluation; 5.1 Introduction; 5.2 Error rate estimation; 5.3 Model selection for MLP models; 5.4 Penalized training; 5.5 Complements and exercises; 6 The Task-based MLP
6.1 Introduction6.2 The task-based MLP; 6.3 Pruning algorithms; 6.4 Interpreting and evaluating task-based MLP models; 6.5 Evaluating the models; 6.6 Conclusion; 6.7 Complements and exercises; 7 Incorporating Spatial Information into an MLP Classifier; 7.1 Allocation and neighbor information; 7.2 Markov random fields; 7.3 Hopfield networks; 7.4 MLP neighbor models; 7.5 Sequential updating; 7.6 Example - MartinTMs farm; 7.7 Conclusion; 7.8 Complements and exercises; 8 Influence Curves for the Multi-layer Perceptron Classifier; 8.1 Introduction; 8.2 Estimators; 8.3 Influence curves
8.4 M-estimators8.5 The MLP; 8.6 Influence curves for pc; 8.7 Summary and Conclusion; 9 The Sensitivity Curves of the MLP Classifier; 9.1 Introduction; 9.2 The sensitivity curve; 9.3 Some experiments; 9.4 Discussion; 9.5 Conclusion; 10 A Robust Fitting Procedure for MLP Models; 10.1 Introduction; 10.2 The effect of a hidden layer; 10.3 Comparison of MLP with robust logistic regression; 10.4 A robust MLP model; 10.5 Diagnostics; 10.6 Conclusion; 10.7 Complements and exercises; 11 Smoothed Weights; 11.1 Introduction; 11.2 MLP models; 11.3 Examples; 11.4 Conclusion
11.5 Cornplernents and exercises12 Translation Invariance; 12.1 Introduction; 12.2 Example 1; 12.3 Example 2; 12.4 Example 3; 12.5 Conclusion; 13 Fixed-slope Training; 13.1 Introduction; 13.2 Strategies; 13.3 Fixing γ or O; 13.4 Example 1; 13.5 Example 2; 13.6 Discussion; Bibliography; Appendix A: Function Minimization; A.l Introduction; A.2 Back-propagation; A.3 Newton-Raphson; A.4 The method of scoring; A.5 Quasi-Newton; A.6 Conjugate gradients; A.7 Scaled conjugate gradients; A.8 Variants on vanilla fiback-propagationfl; A.9 Line search; A.10 The simplex algorithm; A.11 Implementation
A.12 Examples
Record Nr. UNINA-9910830653803321
Dunne Robert A  
Hoboken, N.J. ; ; Chichester, : Wiley, c2007
Materiale a stampa
Lo trovi qui: Univ. Federico II
Opac: Controlla la disponibilità qui
A statistical approach to neural networks for pattern recognition [[electronic resource] /] / Robert A. Dunne
A statistical approach to neural networks for pattern recognition [[electronic resource] /] / Robert A. Dunne
Autore Dunne Robert A
Pubbl/distr/stampa Hoboken, N.J. ; ; Chichester, : Wiley, c2007
Descrizione fisica 1 online resource (289 p.)
Disciplina 006.32
Collana Wiley series in computational statistics
Soggetto topico Perceptrons
Neural networks (Computer science)
ISBN 1-280-93517-0
9786610935178
0-470-14815-2
0-470-14814-4
Formato Materiale a stampa
Livello bibliografico Monografia
Lingua di pubblicazione eng
Nota di contenuto A Statistical Approach to Neural Networks for Pattern Recognition; Contents; Notation and Code Examples; Preface; Acknowledgments; 1 Introduction; 1.1 The perceptron; 2 The Multi-Layer Perceptron Model; 2.1 The multi-layer perceptron (MLP); 2.2 The first and second derivatives; 2.3 Additional hidden layers; 2.4 Classifiers; 2.5 Complements and exercises; 3 Linear Discriminant Analysis; 3.1 An alternative method; 3.2 Example; 3.3 Flexible and penalized LDA; 3.4 Relationship of MLP models to LDA; 3.5 Linear classifiers; 3.6 Complements and exercises; 4 Activation and Penalty Functions
4.1 Introduction4.2 Interpreting outputs as probabilities; 4.3 The fiuniversal approximatorfl and consistency; 4.4 Variance and bias; 4.5 Binary variables and logistic regression; 4.6 MLP models and cross-entropy; 4.7 A derivation of the softmax activation function; 4.8 The finaturalfl pairing and A,; 4.9 A comparison of least squares and cross-entropy; 4.10 Conclusion; 4.11 Complements and exercises; 5 Model Fitting and Evaluation; 5.1 Introduction; 5.2 Error rate estimation; 5.3 Model selection for MLP models; 5.4 Penalized training; 5.5 Complements and exercises; 6 The Task-based MLP
6.1 Introduction6.2 The task-based MLP; 6.3 Pruning algorithms; 6.4 Interpreting and evaluating task-based MLP models; 6.5 Evaluating the models; 6.6 Conclusion; 6.7 Complements and exercises; 7 Incorporating Spatial Information into an MLP Classifier; 7.1 Allocation and neighbor information; 7.2 Markov random fields; 7.3 Hopfield networks; 7.4 MLP neighbor models; 7.5 Sequential updating; 7.6 Example - MartinTMs farm; 7.7 Conclusion; 7.8 Complements and exercises; 8 Influence Curves for the Multi-layer Perceptron Classifier; 8.1 Introduction; 8.2 Estimators; 8.3 Influence curves
8.4 M-estimators8.5 The MLP; 8.6 Influence curves for pc; 8.7 Summary and Conclusion; 9 The Sensitivity Curves of the MLP Classifier; 9.1 Introduction; 9.2 The sensitivity curve; 9.3 Some experiments; 9.4 Discussion; 9.5 Conclusion; 10 A Robust Fitting Procedure for MLP Models; 10.1 Introduction; 10.2 The effect of a hidden layer; 10.3 Comparison of MLP with robust logistic regression; 10.4 A robust MLP model; 10.5 Diagnostics; 10.6 Conclusion; 10.7 Complements and exercises; 11 Smoothed Weights; 11.1 Introduction; 11.2 MLP models; 11.3 Examples; 11.4 Conclusion
11.5 Cornplernents and exercises12 Translation Invariance; 12.1 Introduction; 12.2 Example 1; 12.3 Example 2; 12.4 Example 3; 12.5 Conclusion; 13 Fixed-slope Training; 13.1 Introduction; 13.2 Strategies; 13.3 Fixing γ or O; 13.4 Example 1; 13.5 Example 2; 13.6 Discussion; Bibliography; Appendix A: Function Minimization; A.l Introduction; A.2 Back-propagation; A.3 Newton-Raphson; A.4 The method of scoring; A.5 Quasi-Newton; A.6 Conjugate gradients; A.7 Scaled conjugate gradients; A.8 Variants on vanilla fiback-propagationfl; A.9 Line search; A.10 The simplex algorithm; A.11 Implementation
A.12 Examples
Record Nr. UNINA-9910841024903321
Dunne Robert A  
Hoboken, N.J. ; ; Chichester, : Wiley, c2007
Materiale a stampa
Lo trovi qui: Univ. Federico II
Opac: Controlla la disponibilità qui