(2025/2026) Artificial Neural Networks and Deep Learning
The notes are taken from the books required for the course:
- Ian Goodfellow, Yoshua Bengio, and Aaron Courville, Deep Learning, Editore: MIT Press, ISBN: 978-0262035613, http://www.deeplearningbook.org/.
- Course slides.
You can view/download the PDF here. In the notes folder, you can also see the source code.
For any issue, use the appropriate section.
Course Syllabus
According to the official course syllabus:
- From the Perceptron to Neural Networks and the Feedforward architecture
- Backpropagation and Neural Networks training algorithms, e.g., Adagrad, adam, etc.
- Best practices in neural network training: overfitting and cross-validation, stopping criteria, weight decay, dropout, data resampling and augmentation.
- Image Classification problem and Neural Networks
- Recurrent Neural Networks and other relevant architectures such as (Sparse) Neural Autoencoders
- Theoretical results: Neural Networks as universal approximation tools, vanishing and exploding gradients, etc.
- Introduction to the Deep Learning paradigm and its main differences with respect to classical Machine Learning
- Convolutional Neural Networks (CNN) architecture
- The breakthrough of CNN and their interpretation
- CNN training and data-augmentation
- Structural learning, Long-Short Term Memories, and their applications to text and speech
- Autoencoders and data embedding, word2vec, variational autoencoders
- Transfer Learning for pre-trained Deep models
- Extended models including Fully Convolutional CNN, networks for image segmentation (U-net) and object detection (e.g., R-CNN, YOLO )
- Generative Models (e.g., Generative Adversarial Networks)