(2025/2026) Artificial Neural Networks and Deep Learning
The notes are taken from the books required for the course:
- Ian Goodfellow, Yoshua Bengio, and Aaron Courville, Deep Learning, Editore: MIT Press, ISBN: 978-0262035613, http://www.deeplearningbook.org/.
- Course slides.
You can view/download the PDF here. In the notes folder, you can also see the source code.
For any issue, use the appropriate section.
Correlated Projects
During the Artificial Neural Networks and Deep Learning for HPC course, me and my other three colleagues Alberto Ondei, Abdullah Javed, Filippo Barbari created two projects:
- Time Series Classification. Deep Neural Netowrk (TCN + BiLSTM with Attention) model on multivariate time-series classification.
- Image Classification. Histopathology image classification to predict molecular subtypes. Includes datasets, EDA and preprocessing notebooks, patch extraction, baseline models, and training utilities for quick experiments and reproducible evaluation.
Course Syllabus
According to the official course syllabus:
- From the Perceptron to Neural Networks and the Feedforward architecture
- Backpropagation and Neural Networks training algorithms, e.g., Adagrad, adam, etc.
- Best practices in neural network training: overfitting and cross-validation, stopping criteria, weight decay, dropout, data resampling and augmentation.
- Image Classification problem and Neural Networks
- Recurrent Neural Networks and other relevant architectures such as (Sparse) Neural Autoencoders
- Theoretical results: Neural Networks as universal approximation tools, vanishing and exploding gradients, etc.
- Introduction to the Deep Learning paradigm and its main differences with respect to classical Machine Learning
- Convolutional Neural Networks (CNN) architecture
- The breakthrough of CNN and their interpretation
- CNN training and data-augmentation
- Structural learning, Long-Short Term Memories, and their applications to text and speech
- Autoencoders and data embedding, word2vec, variational autoencoders
- Transfer Learning for pre-trained Deep models
- Extended models including Fully Convolutional CNN, networks for image segmentation (U-net) and object detection (e.g., R-CNN, YOLO )
- Generative Models (e.g., Generative Adversarial Networks)