This repository contains materials for the second half of the "Machine Learning for Sound and Music" course. The course introduces key concepts of machine learning applied to audio and music processing. The materials include lecture notebooks, practical exercises, and examples to help students learn and implement machine learning techniques effectively.
notebooks/: Contains Jupyter notebooks for each lecture.docker/: Includes Docker files to build the image and run the Jupyter server.
- Docker and Docker Compose installed on your machine.
- Basic familiarity with Python and Jupyter Notebooks.
- Recommended: A laptop with a GPU for deep learning tasks.
-
Clone the repository:
git clone <repository-url> cd machine_learning_for_sound_and_music
-
Build the Docker image:
docker-compose build
-
Start the Jupyter server:
docker-compose up
-
Access the Jupyter server: Open your browser and navigate to
http://localhost:8888.
- Lecture 5: Supervised Learning with CNNs. Code for training small MLP and CNN models for a multi-class task.
- Lecture 6: Self-Supervised Learning I: Autoencoders
- Lecture 7: Self-Supervised Representation Learning II: Metric Learning
- Lecture 8: Transformers and the Attention Mechanism
- Lecture 9: Transfer Learning
- Lecture 10: Multi-modal Learning: Text-Audio Models
Contributions are welcome! Please fork the repository and submit a pull request for any improvements or suggestions.
- mirdata: Dataset loader used in the course.
- PyTorch: Preferred library for deep learning tasks.
- Essentia: Library for audio processing.
This repository is licensed under the MIT License. See the LICENSE file for details.