1. | Course Title | Statistical Signal Processing and Statistical Learning | |||||||||||
2. | Code | 4ФЕИТ10025 | |||||||||||
3. | Study program | 10-DPSM, 11-IBS, 12-KIT | |||||||||||
4. | Organizer of the study program (unit, institute, department) | Faculty of Electrical Engineering and Information Technologies | |||||||||||
5. | Degree (first, second, third cycle) | Second cycle | |||||||||||
6. | Academic year/semester | I/1 | 7. | Number of ECTS credits | 6.00 | ||||||||
8. | Lecturer | Dr Venceslav Kafedziski | |||||||||||
9. | Course Prerequisites | ||||||||||||
10. | Course Goals (acquired competencies):
Upon completing the course it is expected that the student will understand and know how to implement the methods and algorithms of statistical signal processing: estimation of parameters, random parameters and random processes, and the methods and algorithms of statistical learning, to know how to apply these methods and algorithms to real life problems and to be capable of researching in the area of statistical signal processing and statistical learning. |
||||||||||||
11. | Course Syllabus:
Random variables and random vectors. Multidimensional Gaussian distribution. Discrete random processes: definition, stationarity and ergodicity, autocorrelation and power spectral density. Parameter estimation: LS, MVUE, ML. Estimation of random parameters: MAP, MMSE, and orthogonality principle. Optimal estimation of discrete random processes: Wiener and Kalman filter. Parametric models of discrete random processes: AR, MA and ARMA. Spectral analysis of discrete random processes: basic methods and high resolution methods. Adaptive signal processing. Array signal processing. Statistical learning. Dimensionality reduction. Discriminative models. Perceptron. Linear regression. SVM. Neural networks. Deep neural networks. Generative models. Bayesian decision theory. Unimodal models. Markov chains. Mixture models: Gaussian mixture models. Hidden Markov models. Bayesian learning. Graphical models. Bayesian networks. Markov random fields. Factor graphs and belief propagation. Applications of the described methods and algorithms. |
||||||||||||
12. | Learning methods:
Lectures, self-learning, term projects, presentations, active participation in the lectures, consultations. |
||||||||||||
13. | Total number of course hours | 180 | |||||||||||
14. | Distribution of course hours | 3 + 3 | |||||||||||
15. | Forms of teaching | 15.1 | Lectures-theoretical teaching | 45 hours | |||||||||
15.2 | Exercises (laboratory, practice classes), seminars, teamwork | 45 hours | |||||||||||
16. | Other course activities | 16.1 | Projects, seminar papers | 30 hours | |||||||||
16.2 | Individual tasks | 30 hours | |||||||||||
16.3 | Homework and self-learning | 30 hours | |||||||||||
17. | Grading | ||||||||||||
17.1 | Exams | 0 points | |||||||||||
17.2 | Seminar work/project (presentation: written and oral) | 50 points | |||||||||||
17.3. | Activity and participation | 0 points | |||||||||||
17.4. | Final exam | 50 points | |||||||||||
18. | Grading criteria (points) | up to 50 points | 5 (five) (F) | ||||||||||
from 51 to 60 points | 6 (six) (E) | ||||||||||||
from 61 to 70 points | 7 (seven) (D) | ||||||||||||
from 71 to 80 points | 8 (eight) (C) | ||||||||||||
from 81 to 90 points | 9 (nine) (B) | ||||||||||||
from 91 to 100 points | 10 (ten) (A) | ||||||||||||
19. | Conditions for acquiring teacher’s signature and for taking final exam | None | |||||||||||
20. | Forms of assessment |
The exam includes a written or oral final exam from the course material listed in the course content and completion and presentation of a term paper/project on a subject mutually agreed by the student and the teacher. |
|||||||||||
21. | Language | Macedonian and English | |||||||||||
22. | Method of monitoring of teaching quality | Self-evaluation. | |||||||||||
23. | Literature | ||||||||||||
23.1. | Required Literature | ||||||||||||
No. | Author | Title | Publisher | Year | |||||||||
1. | H. Jiang | Machine Learning Fundamentals | Cambridge University Press; | 2022 | |||||||||
2. | J. M. Mendel | Lessons in Estimation Theory for Signal Processing, Communications, and Control | Pearson Technology Group | 2008 | |||||||||
23.2. | Additional Literature | ||||||||||||
No. | Author | Title | Publisher | Year | |||||||||
1. | D. G. Manolakis, V. K. Ingle, S. M. Kogon | Statistical and Adaptive Signal Processing: Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing | Artech House | 2005 |