A competition aim at fostering the development of automatic models that can predict human aesthetic ratings of generated songs.
November 15, 2025: We have updated the formula and threshold for calculating Top-Tier Accuracy with detailed information available at the Evaluation section below.
November 10, 2025: We have sent the test set and submission Instructions to all successfully registered teams via email. These are also available in the respective track pages. Kindly note that the final submission deadline is 23:59 November 20, 2025 (AoE time).
With the rapid growth of generative music models, such as song generation (composing melodies, lyrics, harmonies, and vocals), we are entering an exciting new era of personalized music, virtual artists, and multimedia content creation. Despite these advancements, the evaluation of the aesthetic quality of generated music remains a challenge. Traditional metrics like pitch accuracy and signal clarity fall short of capturing the complex emotional and artistic dimensions of music that matter most to listeners. This challenge aims to create a benchmark for assessing the aesthetic quality of automatically generated songs. Participants will develop models that predict human ratings of songs based on musicality, emotional engagement, vocal expressiveness, and overall enjoyment.
Join us to push the boundaries of song aesthetics evaluation and contribute to the future of generative music!
The ICASSP 2026 Automatic Song Aesthetics Evaluation Challenge is designed to foster the development of models that can predict human aesthetic ratings of full-length generated songs. We focus on generating songs that align with human perceptions of musicality, emotional depth, and vocal expressiveness. Participants will be tasked with developing models that predict subjective ratings based on audio inputs.
Objective: Create models that can predict human ratings of aesthetic quality in songs, including dimensions like overall musicality, emotional engagement, and vocal expressiveness.
The competition consists of two tracks:
Track 1: Overall Musicality Score Prediction Participants must predict a single holistic aesthetic score for each song, representing an overall musical impression of the song’s artistic quality.
Track 2: Fine-Grained Aesthetic Dimension Prediction Participants to predict five specific aesthetic dimensions for each song.
Each track will use correlation-based metrics as follows:
We will mesure both system-level and utterance-level.
Quantification Method: Top-Tier Accuracy is uniformly measured using the F1 score.
F1 Score Formula: F1 = 2 × (Precision × Recall) / (Precision + Recall)
Brief Definition: Precision = True Positives / (True Positives + False Positives); Recall = True Positives / (True Positives + False Negatives)
Top-Tier Song Thresholds:
Final results can be submitted using Google form: ICASSP2026 ASAE Challenge Final Submission Form
Each participating team must submit one SCP file per track.
Track 1 submission format:
Each line should contain: utt score
Track 2 submission format:
Each line should contain: utt score1 score2 score3 score4 score5
Please ensure that the utt IDs exactly match those in the provided test sets.
All predicted scores are in a valid numeric format.
Files are named as track1_set1_pred.scp, track1_set2_pred.scp and track2_pred.scp, respectively.
Each team must submit a two-page system description summarizing their method, model architecture, training strategy, and any relevant implementation details.(excluding references)
The format should follow the ICASSP official paper template (available on the ICASSP 2026 website)
Submission of both the prediction files and the system description is required. Missing either will lead to cancellation of the challenge ranking.
The competition provides a baseline system built upon SongEval. The baseline toolkit leverages a trained aesthetic evaluation model on SongEval, enabling automatic scoring of generated songs across five perceptual dimensions, closely aligned with professional musicians’ judgments.
The baseline test validation IDs are available in the val_ids.txt file.
This baseline serves as a reproducible and extensible starting point, helping participants better benchmark their systems and ensuring fair comparison across different approaches.
The challenge is organized by a distinguished team of researchers:
For any inquiries, please contact: Email: yaojx@mail.nwpu.edu.cn
Welcome to join our WeChat group