Loughborough University
Browse

Generating neural architectures from parameter spaces for multi-agent reinforcement learning

Download (1.41 MB)
journal contribution
posted on 2024-08-21, 14:58 authored by Corentin ArtaudCorentin Artaud, Varuna De-SilvaVaruna De-Silva, Rafael PinaRafael Pina, Xiyu ShiXiyu Shi

We explore a data-driven approach to generating neural network parameters to determine whether generative models can capture the underlying distribution of a collection of neural network checkpoints. We compile a dataset of checkpoints from neural networks trained within the multi-agent reinforcement learning framework, thus potentially producing previously unseen combinations of neural network parameters. In particular, our generative model is a conditional transformer-based variational autoencoder that, when provided with random noise and a specified performance metric – in our context, returns – predicts the appropriate distribution over the parameter space to achieve the desired performance metric. Our method successfully generates parameters for a specified optimal return without further fine-tuning. We also show that the parameters generated using this approach are more constrained and less variable and, most importantly, perform on par with those trained directly under the multi-agent reinforcement learning framework. We test our method on the neural network architectures commonly employed in the most advanced state-of-the-art algorithms. 

History

School

  • Loughborough University, London

Published in

Pattern Recognition Letters

Publisher

Elsevier

Version

  • P (Proof)

Rights holder

© The Author(s)

Publisher statement

This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).

Acceptance date

2024-07-13

Publication date

2024-07-20

Copyright date

2024

ISSN

0167-8655

eISSN

1872-7344

Language

  • en

Depositor

Dr Xiyu Shi. Deposit date: 6 August 2024