Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
37th (2023)
Session ID : 2G5-OS-21e-02
Conference information

Scaling Laws of Model Size for World Models
*Makoto SATORyosuke UNNOMasahiro NEGISHIKoudai TABATATaiju WATANABEJunnosuke KAMOHARATaiga KUMERyo OKADAYusuke IWASAWAYutaka MATSUO
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

With the development of deep learning, significant performance improvements have been achieved in computer vision and natural language processing. In these advancements, scaling laws that demonstrate exponential changes in model performance with respect to model size, dataset size, and computational resources used for training have played a significant role. These scaling laws have been reported to hold for various tasks, including image classification, image generation, and natural language processing. However, it has not yet been verified whether these scaling laws are effective for tasks that involve long-horizon predictions. In this study, we investigate the validity of scaling laws for world models from the perspective of model size. We conduct experiments that scale the model sizes of two world models in a video prediction task conditioned on actions using the CARLA dataset, and verify that the loss function decreases exponentially and the scaling law holds when including large-scale autoencoder.

Content from these authors
© 2023 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top