期刊论文详细信息
Semantic web
MIDI2vec: Learning MIDI embeddings for reliable prediction of symbolic music metadata
article
Pasquale Lisena1  Albert Meroño-Peñuela2  Raphaël Troncy1 
[1] EURECOM;King’s College London
关键词: Music;    metadata;    metadata prediction;    graph embeddings;    neural networks;   
DOI  :  10.3233/SW-210446
来源: IOS Press
PDF
【 摘 要 】

An important problem in large symbolic music collections is the low availability of high-quality metadata, which is essential for various information retrieval tasks. Traditionally, systems have addressed this by relying either on costly human annotations or on rule-based systems at a limited scale. Recently, embedding strategies have been exploited for representing latent factors in graphs of connected nodes. In this work, we propose MIDI2vec, a new approach for representing MIDI files as vectors based on graph embedding techniques. Our strategy consists of representing the MIDI data as a graph, including the information about tempo, time signature, programs and notes. Next, we run and optimise node2vec for generating embeddings using random walks in the graph. We demonstrate that the resulting vectors can successfully be employed for predicting the musical genre and other metadata such as the composer, the instrument or the movement. In particular, we conduct experiments using those vectors as input to a Feed-Forward Neural Network and we report good comparable accuracy scores in the prediction with respect to other approaches relying purely on symbolic music, avoiding feature engineering and producing highly scalable and reusable models with low dimensionality. Our proposal has real-world applications in automated metadata tagging for symbolic music, for example in digital libraries for musicology, datasets for machine learning, and knowledge graph completion.

【 授权许可】

Unknown   

【 预 览 】
附件列表
Files Size Format View
RO202307140004908ZK.pdf 931KB PDF download
  文献评价指标  
  下载次数:5次 浏览次数:1次