This paper presents a novel approach to neural instrument sound synthesis using a two-stage semi-supervised learning framework capable of generating pitch-accurate, high-quality music samples from an expressive timbre latent space. Existing approaches that achieve sufficient quality for music production often rely on high-dimensional latent representations that are difficult to navigate and provide unintuitive user experiences. We address this limitation through a two-stage training paradigm: first, we train a pitch-timbre disentangled 2D representation of audio samples using a Variational Autoencoder, second, we use this representation as conditioning input for a Transformer-based generative model. The learned 2D latent space serves as an intuitive interface for navigating and exploring the sound landscape. We demonstrate that the proposed method effectively learns a disentangled timbre space, enabling expressive and controllable audio generation with reliable pitch conditioning. Experimental results show the model's ability to capture subtle variations in timbre while maintaining a high degree of pitch accuracy. The usability of our method is demonstrated in this interactive web application, highlighting its potential as a step towards future music production environments that are both intuitive and creatively empowering.
Our interactive interface lets you generate musical samples by selecting points on a 2D plane. To select a pitch for generation, you have two options: 1. use the slider to select the desired note or 2. use your computer keyboard, which is mapped to a MIDI piano starting with note C3 on the 'a' key. The 'q' key allows you to toggle between two octaves. A fraction of the training data is displayed for orientation. Click on a certain location to play the related sample (speaker or headphones recommended).
This schematic depicts the training procedure of our model. In the first stage, a VAE with a 2D latent bottleneck is trained. In the second stage, the Transformer model is trained with the VAE as a conditioning model.
@inproceedings{DAFx25_paper_58,
author = "Limberg, Christian and Schulz, Fares and Zhang, Zhe and Weinzierl, Stefan",
title = "{Pitch-Conditioned Instrument Sound Synthesis From an Interactive Timbre Latent Space}",
booktitle = "Proceedings of the 28-th Int. Conf. on Digital Audio Effects (DAFx25)",
editor = "Gabrielli, L. and Cecchi, S.",
location = "Ancona, Italy",
eventdate = "2025-09-02/2025-09-05",
year = "2025",
month = "Sept",
publisher = "",
issn = "2413-6689",
doi = "",
pages = ""
}