Category Programme announcements

AIM at AES AIMLA Conference

AIMLA conference logoThe AES International Conference on Artificial Intelligence and Machine Learning for Audio (AIMLA 2025) will be hosted by the Centre for Digital Music of Queen Mary University of London and is taking place on Sept. 8-10, 2025.

Several AIM members are involved in the organisation of the conference, including but not limited to:

  • Soumya Vanka (Special Sessions Co-Chair)
  • Franco Caspe (Special Sessions Co-Chair)
  • Farida Yusuf (Sponsorship Chair)
  • Chin-Yun Yu (Late Breaking Papers Chair)

Several papers and presentations will be made from AIM members at AIMLA as well. The following peer-reviewed papers will be presented at the conference:

The following late-breaking poster from AIM members will be presented at AIMLA:

Last but not least, the following tutorial will be co-presented by AIM member Franco Caspe:

See you in London!


Call for Late Breaking Demo Papers: First AES International Conference on AI and Machine Learning for Audio (AIMLA 2025)

AIMLA conference logoThe AES International Conference on Artificial Intelligence and Machine Learning for Audio (AIMLA 2025), hosted at the Centre for Digital Music of Queen Mary University of London and taking place on Sept. 8-10, 2025 is calling for Late Breaking Demo Paper submissions.

We are seeking 2-page extended abstracts showcasing prototype systems and early research results that are highly relevant to the conference theme. At least one author must register for the conference and present their work as a poster in person during the main track poster session.

Submissions open on July 1, 2025 with a deadline of August 1, 2025. Submissions will be reviewed on a rolling basis.

For more information on submission guidelines, templates, and technical requirements, please visit: https://aes2.org/contributions/2025-1st-aes-international-conference-on-artificial-intelligence-and-machine-learning-for-audio-call-for-contributions/


AIM involvement in MIREX 2025

mirex logoMIREX (Music Information Retrieval Evaluation eXchange) is a prominent evaluation platform in the field of music information retrieval. Researchers are invited to submit novel algorithms for a variety of music-related tasks and receive standardized evaluation results, with the opportunity to present posters during the annual ISMIR conference. For more details on submission, please see the following link:

This year, AIM PhD students Yinghao Ma and Huan Zhang introduced new tasks to the platform, including Multimodal Music QA, Expressive Piano Performance Rendering, alongside traditional MIR challenges and emerging understanding/generation tasks. Specifically, we are coordinating the following tasks:

Music Reasoning QA
Task Captain: Yinghao Ma
The MIREX 2025 Music Reasoning Question Answering (QA) Task challenges participants to develop models capable of answering natural language questions that require understanding and reasoning over musical audio. This task seeks to advance the frontier of machine music intelligence by evaluating models on their ability to reason about all kinds of music information musical structure, instrument presence, melody information, vocal content, and environmental context etc., along with knowledge in music theory and music history.
Participants will build systems that answer multiple-choice questions grounded in audio inputs. The task includes questions from four curated subsets (Music, Music-Speech, Sound-Music, Sound-Music-Speech) from the MMAR benchmark, and Music-subset with image caption from the OmniBench benchmark. Each question is paired with an audio clip and 2-4 different choices.

RenCon: Expressive Piano Performance Rendering Contest
Task Captain: Huan Zhang
Expressive Performance Rendering (https://ren-con2025.vercel.app/) is a task that challenges participants to develop systems capable of rendering expressive musical performances from symbolic scores in MusicXML format. We accept system that generate symbolic (MIDI) or audio (wav) renderings, and the output shall contain human-like expressive deviation from the MusicXML score.
Similar to AI song contest, the evaluation of expressive rendering is subjective and requires human judges to assess. Thus, we have a two-phase competition structure: Phase 1 – Preliminary Round (Online) Submit performances of assigned and free-choice pieces. The submission period is open from May 30, 2025 to Aug 20, 2025. After the submission deadline, the preliminary round page will be finalized with the list of participants and their submissions, and the online evaluation will take place. Phase 2 – Live Contest at ISMIR (Daejeon, Korea) Top systems from preliminary round will be invited to render a surprise piece live at ISMIR, using their system in real time. The live contest is open to all ISMIR attendees, as well as the general public. The audience will be able to listen to the live performances and vote for their favorite system.

Audio Beat Tracking
Task Captain: Wenye Ma & Yinghao Ma
The aim of the automatic beat tracking task is to track each beat locations in a collection of sound files. Unlike the Audio Tempo Extraction task, which aim is to detect tempi for each file, the beat tracking task aims at detecting all beat locations in recordings. The algorithms will be evaluated in terms of their accuracy in predicting beat locations annotated by a group of listeners.

Audio Key Detection
Task Captain: Wenye Ma & Yinghao Ma
Audio Key Detection aims to identify the musical key (e.g., C major, A minor) of an audio recording. This involves determining both the tonic (root pitch) and the mode (major or minor) from the audio signal.