RGTU/RGPV EC-7202 Neural Networks Syllabus
RGTU/RGPV Neural Networks SYLLABUS
Electronics and Communication Engineering EC 7th Semester Syllabus
Unit-II : Perceptrons
Single layer perceptrons: adaptive filtering problem, unconstrained optimization technique, linear least squares filter, least mean square algorithm (LMS), perceptron convergence theorem
Multi layer perceptron: architecture, back propagation algorithm, generalization, approximations of functions, network pruning techniques
Unit-IV : Information- Theoretic Models
Entropy, maximum entropy principle, mutual information, Kullback-Leibler divergence, mutual information as an objective function to be optimized, maximum mutual information principle, infomax and redundancy reduction, spatially coherent and incoherent features, independent components analysis, maximum likelihood estimation, maximum entropy method.
Unit V : Dynamically Driven Recurrent Networks
introduction, recurrent network architectures, state space model, non-linear autogressive with exogenous inputs model, computational power of recurrent networks, learning algorithms, back propagation through time, real time recurrent learning, Kalman filter, decoupled Kalman filter, vanishing gradients in recurrent networks, system identification, model reference adaptive control.
References:
1. S. Haykin: Neural Networks- A Comprehensive Foundation, PHI Learning.
2. S. N. Sivanandam, S. Sumathi and S. N. Deepa: Introduction to Neural Networks using Matlab 6.0, TMH, New Delhi.
3. J. A Freeman and D. M. Skapura: Fundamentals of Neural Networks- algorithms, applications and programming techniques, Pearson Education.
4. M. T. Hagan, H. B. Demuth and M. Beale: Neural Network Design, Cengage Learning.
5. J.A Anderson: An introduction ro Neural Networks, PHI Learning.
6. Satish Kumar: Neural Networks, TMH, New Delhi.
EC-7202 Neural Networks Course Content:
Unit-I : Neural Network (NN)
Introduction, benefits of neural network, models of a neuron, neural network as directed graph, network architectures, artificial intelligence and neural network.
Learning processes: error correction learning, memory based learning, Hebbian learning, competitive learning, Boltzman learning, learning tasks, adaptation, statistical nature of learning process, statistical learning theory.
Introduction, benefits of neural network, models of a neuron, neural network as directed graph, network architectures, artificial intelligence and neural network.
Learning processes: error correction learning, memory based learning, Hebbian learning, competitive learning, Boltzman learning, learning tasks, adaptation, statistical nature of learning process, statistical learning theory.
Unit-II : Perceptrons
Single layer perceptrons: adaptive filtering problem, unconstrained optimization technique, linear least squares filter, least mean square algorithm (LMS), perceptron convergence theorem
Multi layer perceptron: architecture, back propagation algorithm, generalization, approximations of functions, network pruning techniques
.
Unit-III : Radial Basis Function (RBF) Networks
Cover’s theorem on the separability of patterns, interpolation problem, supervised learning as an Ill-posed hypersurface reconstruction problem, regularization theory, regularization network, generalized radial basis function networks (RBF), estimation of the regularization parameter, approximation properties of RBF networks, comparison of RBF networks and multilayer perceptrons, Kernel regression and its relation to RBF networks, learning strategies.
Unit-III : Radial Basis Function (RBF) Networks
Cover’s theorem on the separability of patterns, interpolation problem, supervised learning as an Ill-posed hypersurface reconstruction problem, regularization theory, regularization network, generalized radial basis function networks (RBF), estimation of the regularization parameter, approximation properties of RBF networks, comparison of RBF networks and multilayer perceptrons, Kernel regression and its relation to RBF networks, learning strategies.
Unit-IV : Information- Theoretic Models
Entropy, maximum entropy principle, mutual information, Kullback-Leibler divergence, mutual information as an objective function to be optimized, maximum mutual information principle, infomax and redundancy reduction, spatially coherent and incoherent features, independent components analysis, maximum likelihood estimation, maximum entropy method.
Unit V : Dynamically Driven Recurrent Networks
introduction, recurrent network architectures, state space model, non-linear autogressive with exogenous inputs model, computational power of recurrent networks, learning algorithms, back propagation through time, real time recurrent learning, Kalman filter, decoupled Kalman filter, vanishing gradients in recurrent networks, system identification, model reference adaptive control.
References:
1. S. Haykin: Neural Networks- A Comprehensive Foundation, PHI Learning.
2. S. N. Sivanandam, S. Sumathi and S. N. Deepa: Introduction to Neural Networks using Matlab 6.0, TMH, New Delhi.
3. J. A Freeman and D. M. Skapura: Fundamentals of Neural Networks- algorithms, applications and programming techniques, Pearson Education.
4. M. T. Hagan, H. B. Demuth and M. Beale: Neural Network Design, Cengage Learning.
5. J.A Anderson: An introduction ro Neural Networks, PHI Learning.
6. Satish Kumar: Neural Networks, TMH, New Delhi.
ConversionConversion EmoticonEmoticon