FPF:UINA500 Deep Learning - Course Information
UINA500 Deep Learning
Faculty of Philosophy and Science in OpavaWinter 2024
- Extent and Intensity
- 2/2/0. 6 credit(s). Type of Completion: zk (examination).
- Teacher(s)
- doc. Ing. Petr Sosík, Dr. (lecturer)
Mgr. Tomáš Filip (seminar tutor)
doc. Ing. Petr Sosík, Dr. (seminar tutor) - Guaranteed by
- doc. Ing. Petr Sosík, Dr.
Institute of Computer Science – Faculty of Philosophy and Science in Opava - Timetable
- Wed 10:35–12:10 232
- Timetable of Seminar Groups:
- Course Enrolment Limitations
- The course is also offered to the students of the fields other than those the course is directly associated with.
- fields of study / plans the course is directly associated with
- Computer Science (programme FPF, CompSci-np)
- Course objectives
- This classical branch of Artificial Intelligence covers a range of machine learning algorithm, typically benefiting from gradient-based learning methods. The most typical learning model is the artificial neural net with many efficient algorithms capable to learn from examples, to generalize knowledge and to search for approximate solutions of intractable problems. These algorithms can be run on special parallel machines but also on classical computers.
- Learning outcomes
- The student will get acquainted with the basic mathematical and structural principles of deep learning. Will be able to design and test deep learning networks for a variety of tasks such as classification, image analysis, comprehension and text generation, or strategic decision making.
- Syllabus
- 1. The structure of biological neuron, mathematical model of a simple neuron and a multi-layer neural network. Advantages and applications of artificial neural nets in deep learning. 2. Active, adaptive and organization dynamics, neural training schemes (supervised/unsupervised/reinforcement). Training and testing sets, training process, the overfitting problem. 3. The perceptron and its training algorithm. Implementation of simple logic functions. Limited capabilities of single-layer perceptron. 4. Multilayer networks and the Backpropagation (BP) algorithm. Modifications and improvements of the BP algorithm, conjugate-gradient methods, resilient propagation, further training methods. 5. Deep feedforward networks, architecture design, regularization methods. 6. Optimization of deep learning, parameter initialization methods, adaptive learning rates, meta-algorithms. 7. Deep learning in recurrent networks, topology and training algorithms, recursive networks. Problem of long dependencies, LSTM networks and related models. 8. Radial Basis Function networks, organization and active dynamics. Three phases of training, properties, applications, a comparison with multilayer perceptron. 9. Competitive networks and vector quantization problem, Lloyd's algorithm. The Kohonen training rule, the UCL, DCL a SCL algorithms. Self-organizing maps – SOM. The ART networks, principles and properties.
- Literature
- required literature
- Chollet, Francois. Deep learning with Python. Simon and Schuster, 2017.
- GOODFELLOW, Ian, BENGIO, Yoshua, COURVILLE, Aaron. Deep Learning. Cambridge, Massachusetts: MIT Press, 2016
- recommended literature
- MIRJALILI, Seyedali. Evolutionary Algorithms and Neural Networks: Theory and Applications. New York, NY: Springer International Publishing, 2018.
- Teaching methods
- Interactive lecture Lecture with a video analysis
- Assessment methods
- Individual projects and exercises for solutions at home.
- Language of instruction
- English
- Enrolment Statistics (recent)
- Permalink: https://is.slu.cz/course/fpf/winter2024/UINA500