Machine Learning for Speech and Audio Processing

Lecturer: Prof. Dr.-Ing. Peter Jax

Contact: Lars Thieling, Maximilian Kentgens

Type: Master lecture

Credits: 4

Lecture in RWTHonline
Exercise in RWTHonline
Learning room RWTHmoodle
(Registration via RWTHonline)

Course language: English

Lecture slides and Exercise problems will be published in RWTHmoodle.



from Friday, April 8, 2022
08:30 - 10:00
Lecture Room 4G


from Friday, April 8, 2022
10:15 - 11:00
Lecture Room 4G

Consultation hours:

Dates will be announced.


Thursday, August 11, 2022
Lecture room PPS H1/H2

Exam duration: 90 minutes

Remarks: The exam is in written form. The date corresponds also to the proof of achievements (Leistungsachweise) in written form.

Resources: You are allowed to bring one hand-written DIN A4 formula sheet (front and back). Any other written material (e.g., lecture notes, exercise notes) is not allowed. A non-programmable calculator is allowed.

Please note: Please bring along your student ID (BlueCard)!


Tuesday, September 20, 2022
Lecture Room 4G

The lecture "Machine Learning for Speech and Audio Processing (MLSAP)" addresses especially students of the Master's program "Electrical Engineering, Information Technology and Computer Engineering". The formal connection to the module catalogs can be found at RWTHonline.


In this one term lecture the fundamental methods of machine learning with applications to problems in speech and audio signal processing are presented:

  • Fundamentals of Classification and Estimation
    • Basic Problems of Classification
    • Feature Extraction Techniques
    • Basic Classification Schemes
  • Probabilistic Models
    • Stochastic Processes and Models
    • Gaussian Mixture Models (GMMs)
    • Hidden Markov Models (HMMs)
    • Training Methods
    • Bayesian Probability Theory: Classification and Estimation
    • Particle Filter
  • Non-Negative Matrix Factorization (NMF)
    • Dictionary-based concept
  • Neural Network and Deep Learning
    • Feed-Forward Neural Networks
    • Fundamental Applications
    • Learning Strategies: Supervised vs Unsupervised vs Reinforcement Learning
    • Training of Synaptic Weights: Backpropagation and Stochastic Gradient Descent
    • Behavior of Learning and the “Magic” of Setting Hyper‐Parameters
    • Generative Networks as a Complement to Directed Graphs
    • From „Shallow“ to „Deep“: Trade Comprehensibility for Performance
    • Specific Network Architectures
    • Applications in Signal Processsing
    • Interpretations and Realizations

Exercises are offered to gain a deeper understanding on the basis of practical examples.


The results of the evaluation are summarized below.

Summer term 2020

Participants of the evaluation 5

Global grade: 1,6
Concept of the lecture: 1,5
Instruction and behaviour: 1,6