Multimodal AI

CMI-RewardBench: Evaluating Music Reward Models with Compositional Multimodal Instruction

YYinghao MaHHaiwen XiaHHewei GaoWWeixiong ChenYYuxin YeYYuchen YangSSungkyun ChangMMingshuo DingYYizhi LiRRuibin YuanSSimon DixonEEmmanouil Benetos
Published
February 28, 2026
Authors
12
Word Count
17,843
Code
Includes code

Unified reward model benchmark for evaluating music generation across text, lyrics, and audio instructions.

Abstract

While music generation models have evolved to handle complex multimodal inputs mixing text, lyrics, and reference audio, evaluation mechanisms have lagged behind. In this paper, we bridge this critical gap by establishing a comprehensive ecosystem for music reward modeling under Compositional Multimodal Instruction (CMI), where the generated music may be conditioned on text descriptions, lyrics, and audio prompts. We first introduce CMI-Pref-Pseudo, a large-scale preference dataset comprising 110k pseudo-labeled samples, and CMI-Pref, a high-quality, human-annotated corpus tailored for fine-grained alignment tasks. To unify the evaluation landscape, we propose CMI-RewardBench, a unified benchmark that evaluates music reward models on heterogeneous samples across musicality, text-music alignment, and compositional instruction alignment. Leveraging these resources, we develop CMI reward models (CMI-RMs), a parameter-efficient reward model family capable of processing heterogeneous inputs. We evaluate their correlation with human judgments scores on musicality and alignment on CMI-Pref along with previous datasets. Further experiments demonstrate that CMI-RM not only correlates strongly with human judgments, but also enables effective inference-time scaling via top-k filtering. The necessary training data, benchmarks, and reward models are publicly available.

Key Takeaways

  • 1

    CMI-RewardBench unifies fragmented music evaluation tools into a single benchmark for compositional multimodal instructions.

  • 2

    The CMI-RM model achieves parameter efficiency while handling text, lyrics, and audio inputs simultaneously for music generation.

  • 3

    Position-consistency filtering in dataset creation improved reliability by validating LLM preferences across reversed input orderings.

Limitations

  • Existing reward models struggle with flexible multimodal inputs, showing capability gaps even with state-of-the-art multimodal LLMs.

  • Traditional metrics like FAD operate at distribution level, failing to provide sample-level signals necessary for training alignment.

Keywords

music reward modelingCompositional Multimodal Instructionpreference datasetpseudo-labeled sampleshuman-annotated corpusunified benchmarkreward modelsparameter-efficienttop-k filtering

More in Multimodal AI

View all
CMI-RewardBench: Evaluating Music Reward Models with Compositional Multimodal Instruction | Paperchime