An adaptive music generation architecture for games based on the deep learning Transformer model - Sorbonne Université Access content directly
Preprints, Working Papers, ... Year : 2022

An adaptive music generation architecture for games based on the deep learning Transformer model

Abstract

This paper presents an architecture for generating music for video games based on the Transformer deep learning model. The system generates music in various layers, following the standard layering strategy currently used by composers designing video game music. The music is adaptive to the psychological context of the player, according to the arousal-valence model. Our motivation is to customize music according to the player's tastes, who can select his preferred style of music through a set of training examples of music. We discuss current limitations and prospects for the future, such as collaborative and interactive control of the musical components.
Fichier principal
Vignette du fichier
2207.01698.pdf (566.84 Ko) Télécharger le fichier
Origin Files produced by the author(s)

Dates and versions

hal-03715919 , version 1 (06-07-2022)

Identifiers

  • HAL Id : hal-03715919 , version 1

Cite

Gustavo Amaral Costa, Augusto Baffa, Jean-Pierre Briot, Bruno Feijo, Antonio Luz Furtado. An adaptive music generation architecture for games based on the deep learning Transformer model. 2022. ⟨hal-03715919⟩
57 View
114 Download

Share

Gmail Mastodon Facebook X LinkedIn More