Les médias liés à cet évènement

Hybrid deep learning for music analysis and synthesis - Gaël Richard

16 novembre 2023 53 min

Invariance learning for a music indexing robust to sound modifications - Rémi Mignot

16 novembre 2023 51 min

GDR ISIS, Méthodes et modèles en traitement de signal, Introduction

16 novembre 2023 05 min

Labeling a Large Music Catalog - Romain Hennequin

16 novembre 2023 01 h 04 min

Introduction à la journée d'études du GdR IASIS dédiée à la synthèse audio - Thomas Hélie, Mathieu Lagrange

7 novembre 2024

Audio Language Models - Neil Zeghidour

7 novembre 2024

Poster sessions - Clara Boukhemia, Samir Sadok, Amandine Brunetto, Haoran Sun, Vincent Lostanlen, Morgane Buisson, Xiran Zhang, Reyhaneh Abbasi, Ainė Drėlingytė, Étienne Paul André, Yuexuan Kong, Étienne Bost, Axel Marmoret, Javier Nistal, Hugo Pauget Ballesteros

7 novembre 2024

AI in 64Kbps: Lightweight neural audio synthesis for embedded instruments - Philippe Esling

7 novembre 2024

Music sound synthesis using machine learning - Fanny Roche

7 novembre 2024

Grey-box modelling informed by physics: Application to commercial digital audio effects - Judy Najnudel

7 novembre 2024

Hybrid deep learning for music analysis and synthesis - Gaël Richard

16 novembre 2023 53 min

Invariance learning for a music indexing robust to sound modifications - Rémi Mignot

16 novembre 2023 51 min

GDR ISIS, Méthodes et modèles en traitement de signal, Introduction

16 novembre 2023 05 min

Labeling a Large Music Catalog - Romain Hennequin

16 novembre 2023 01 h 04 min

Basic Pitch: A lightweight model for multi-pitch, note and pitch bend estimations in polyphonic music

0:00/0:00

"Basic-pitch" is a lightweight neural network for musical instrument transcription, which supports polyphonic outputs and generalizes to a wide variety of instruments (including vocals). In this talk, we will discuss how we built and evaluated this efficient and simple model, which experimentally showed to be substantially better than a comparable baseline in detecting notes. The model is trained to jointly predict frame-wise onsets, multi-pitch and note activations, and we experimentally showed that this multi-output structure improves the resulting frame-level note accuracy. We will also listen to examples using (and misusing) this model for creative purposes, using our open-source python library, or demo website: thanks to its scalability, the model can run on the browser, and your audio doesn't even leave your own computer.

Paper: https://arxiv.org/abs/2203.09893
Code: https://github.com/spotify/basic-pitch
Demo: https://basicpitch.spotify.com

intervenants

informations

Type
Séminaire / Conférence
Lieu de représentation
Ircam, Salle Igor-Stravinsky (Paris)
durée
43 min
date
16 novembre 2023

IRCAM

1, place Igor-Stravinsky
75004 Paris
+33 1 44 78 48 43

heures d'ouverture

Du lundi au vendredi de 9h30 à 19h
Fermé le samedi et le dimanche

accès en transports

Hôtel de Ville, Rambuteau, Châtelet, Les Halles

Institut de Recherche et de Coordination Acoustique/Musique

Copyright © 2022 Ircam. All rights reserved.