information

Type
Séminaire / Conférence
performance location
Ircam, Salle Igor-Stravinsky (Paris)
date
November 7, 2024

One of the major challenges of the synthesizer market and sound synthesis today lies in proposing new forms of synthesis allowing the creation of brand new sonorities while offering musicians more intuitive and perceptually meaningful control to help them find the perfect sound more easily. Indeed, today's synthesizers are very powerful tools offering musicians a wide range of possibilities for creating sound textures, but the control of parameters still lacks user-friendliness and generally requires expert knowledge to manipulate. This presentation will focus on machine learning methods for sound synthesis, enabling the generation of new, high-quality sounds while providing perceptually relevant control parameters.

In a first part of this talk, we will focus on the perceptual characterization of synthetic musical timbre by highlighting a set of verbal descriptors frequently and consensually used by musicians. Secondly, we will explore the use of machine learning algorithms for sound synthesis, and in particular different models of the "autoencoder" type, for which we have carried out an in-depth comparative study on two different datasets. Then, this presentation will focus on the perceptual regularization of the proposed model, based on the perceptual characterization of synthetic timbre presented in the first part, to enable (at least partial) perceptually relevant control of sound synthesis. Finally, in the last part of this talk, we will quickly present some of the latest tests we conducted using more recent neural synthesis models.

speakers

From the same archive

Introduction à la journée d'études du GdR IASIS dédiée à la synthèse audio - Thomas Hélie, Mathieu Lagrange

Cette journée d'études dédiée à la synthèse audio est organisée dans le cadre de l'axe « Audio, Vision, Perception » du GdR IASIS. À cette occasion, nous invitons quatre orateurs et oratrices, issu-e-s de la recherche française publique et

November 7, 2024

Video

Audio Language Models - Neil Zeghidour

Audio analysis and audio synthesis require modeling long-term, complex phenomena and have historically been tackled in an asymmetric fashion, with specific analysis models that differ from their synthesis counterpart. In this presentation,

November 7, 2024

Video

Poster sessions - Clara Boukhemia, Samir Sadok, Amandine Brunetto, Haoran Sun, Vincent Lostanlen, Morgane Buisson, Xiran Zhang, Reyhaneh Abbasi, Ainė Drėlingytė, Étienne Paul André, Yuexuan Kong, Étienne Bost, Axel Marmoret, Javier Nistal, Hugo Pauget Ballesteros

November 7, 2024

Video

AI in 64Kbps: Lightweight neural audio synthesis for embedded instruments - Philippe Esling

The research project led by the ACIDS group at IRCAM aims to model musical creativity by extending probabilistic learning approaches to the use of multivariate and multimodal time series. Our main object of study lies in the properties and

November 7, 2024

Video

Grey-box modelling informed by physics: Application to commercial digital audio effects - Judy Najnudel

In an ever-expanding and competitive market, commercial digital audio effects have significant constraints. Their computation load must be reduced so that they can be operable in real-time. They must be easily controllable through parameter

November 7, 2024

Video

share


Do you notice a mistake?

IRCAM

1, place Igor-Stravinsky
75004 Paris
+33 1 44 78 48 43

opening times

Monday through Friday 9:30am-7pm
Closed Saturday and Sunday

subway access

Hôtel de Ville, Rambuteau, Châtelet, Les Halles

Institut de Recherche et de Coordination Acoustique/Musique

Copyright © 2022 Ircam. All rights reserved.