Course Repository | TU Berlin | Audio Communication Group
Instructor: Fares Schulz (fares.schulz@tu-berlin.de)
Teaching Assistant: Lina Campanella (l.campanella@tu-berlin.de)
This repository contains all learning resources for Deep Learning for Audio Data. The seminar covers the mathematical and algorithmic fundamentals of machine learning, with particular emphasis on deep learning and neural networks for audio data. Core topics include:
- Mathematical and algorithmic foundations of machine learning, deep learning, and neural networks
- Data exploration, preprocessing, and augmentation techniques for audio
- Feature extraction and representation learning for audio signals
- Machine learning models and architectures tailored to audio applications
- Training and evaluation strategies, including relevant metrics and validation approaches
- Model export, deployment, and inference for production environments
- Deep learning workflows, pipelines, and lifecycles – setup, optimization, and best practices
A special emphasis is placed on creative applications of deep learning in the music production domain, exploring how these techniques can be applied to music generation, sound design, audio effects, and other artistic contexts.
You can find more information about the course on MOSES.
dl4ad/
├── slides/ # Lecture slides (Reveal.js presentations)
│ ├── 01-introduction.md
│ ├── 02-history.md
│ ├── 03-machine_learning_fundamentals.md
│ ├── ... (more slides)
│ └── assets/ # Images, videos, fonts for slides
│
├── notebooks/ # Jupyter notebooks for practical exercises
│ ── 01_introduction.ipynb
│ ── 02_machine_learning_fundamentals.ipynb
│ ── ... (more notebooks)
│
├── extras/ # Additional learning resources
│ ├── machine_learning_history.md
│ ├── neural_audio_systems_milestones.md
│ └── animations/ # Manim animations source code
│
├── scripts/ # Development scripts
│ ├── build.js # Build slides
│ └── dev-server.js # Development server with hot reload
│
├── requirements.txt # Python dependencies
├── package.json # Node.js dependencies for slides
└── README.md # This file
- Python 3.8+ (recommended: Python 3.13)
- Node.js 18+ (for viewing slides locally)
- Git
- VS Code (recommended IDE, but any IDE works)
git clone https://github.com/faressc/dl4ad.git
cd dl4adCreate and activate a virtual environment:
macOS/Linux:
python -m venv .venv
source .venv/bin/activateWindows:
python -m venv .venv
.venv\Scripts\activateInstall Python dependencies:
pip install -r requirements.txt(Optional) Install the requirements for generating animations:
pip install -r requirements_animations.txtInstall Node.js dependencies:
npm install- Open VS Code in the repository directory
- Open a notebook from the
notebooks/folder - Select Python interpreter from your
.venvenvironment - Run cells to execute code
The latest slides are hosted online at: https://faressc.github.io/dl4ad/
Start the development server with hot reload:
npm start
# or
npm run devThen open your browser to http://localhost:8080
Build static slides:
npm run buildThe built slides will be in the dist/ folder.
This repository uses Manim for creating mathematical animations.
Generate animations:
manim extras/animations/03-machine_learning_fundamentals.pyConfiguration is in manim.cfg.
This repository is licensed under the Creative Commons Attribution-NonCommercial 4.0 International License.
Course developed by the Computer Music and Neural Audio Systems Research Team
Audio Communication Group
Technische Universität Berlin
The following individuals contributed significantly to the course materials:
- Fares Schulz
- Lina Campanella (Jupyter notebooks)
For more information about our research group, visit the Audio Communication Group website.
This course has been inspired by the following excellent course: