pwshub.com

New AI model “learns” how to simulate Super Mario Bros. from video footage

At first glance, these AI-generated <em>Super Mario Bros.</em> videos are pretty impressive. The more you watch, though, the more glitches you'll see.

At first glance, these AI-generated Super Mario Bros. videos are pretty impressive. The more you watch, though, the more glitches you'll see.

Last month, Google's GameNGen AI model showed that generalized image diffusion techniques can be used to generate a passable, playable version of Doom. Now, researchers are using some similar techniques with a model called MarioVGG to see if an AI model can generate plausible video of Super Mario Bros. in response to user inputs.

The results of the MarioVGG model—available as a pre-print paper published by the crypto-adjacent AI company Virtuals Protocol—still display a lot of apparent glitches, and it's too slow for anything approaching real-time gameplay at the moment. But the results show how even a limited model can infer some impressive physics and gameplay dynamics just from studying a bit of video and input data.

The researchers hope this represents a first step toward "producing and demonstrating a reliable and controllable video game generator," or possibly even "replacing game development and game engines completely using video generation models" in the future.

Watching 737,000 frames of Mario

To train their model, the MarioVGG researchers (GitHub users erniechew and Brian Lim are listed as contributors) started with a public data set of Super Mario Bros. gameplay containing 280 "levels'" worth of input and image data arranged for machine-learning purposes (level 1-1 was removed from the training data so images from it could be used in the evaluation). The more than 737,000 individual frames in that data set were "preprocessed" into 35 frame chunks so the model could start to learn what the immediate results of various inputs generally looked like.

To "simplify the gameplay situation," the researchers decided to focus only on two potential inputs in the data set: "run right" and "run right and jump." Even this limited movement set presented some difficulties for the machine-learning system, though, since the preprocessor had to look backward for a few frames before a jump to figure out if and when the "run" started. Any jumps that included mid-air adjustments (i.e., the "left" button) also had to be thrown out because "this would introduce noise to the training dataset," the researchers write.

  • MarioVGG takes a single gameplay frame and a text input action to generate multiple video frames.

  • The last frame of a generated video sequence can be used as the baseline for the next set of frames in the video.

  • The AI-generated arc of Mario's jump is pretty accurate (even as the algorithm creates random obstacles as the screen "scrolls").

  • MarioVGG was able to infer the physics of behaviors like running off a ledge or running into obstacles.

  • A particularly bad example of a glitch that causes Mario to simply disappear from the scene at points.

After preprocessing (and about 48 hours of training on a single RTX 4090 graphics card), the researchers used a standard convolution and denoising process to generate new frames of video from a static starting game image and a text input (either "run" or "jump" in this limited case). While these generated sequences only last for a few frames, the last frame of one sequence can be used as the first of a new sequence, feasibly creating gameplay videos of any length that still show "coherent and consistent gameplay," according to the researchers.

Source: arstechnica.com

Related stories
3 days ago - With Google, Microsoft, Apple and OpenAI creating new AI models and systems, it's good to stay up to date on all the latest terminology.
1 week ago - AI hype is in full swing, with companies like Google, Microsoft, Meta and Apple putting it in everything. There's a lot of new words being thrown around. This glossary is your one-stop shop.
1 day ago - Anthropic Claude 3.5 shines in Chatterbox Labs safety test Feature  Anthropic has positioned itself as a leader in AI safety, and in a recent analysis by Chatterbox Labs, that proved to be the case.…
1 month ago - Apple's generative AI plans have only been shared in recent months. It's been leaning on artificial intelligence since years before the iPhone.
3 weeks ago - Isn't the education system in enough trouble already? The UK government is set to equip teachers with AI tools to help them "mark and plan lessons."…
Other stories
2 minutes ago - Many left reeling from July's IT meltdown, but not to worry, it was all unavoidable Germany's Federal Office for Information Security (BSI) says one in ten organizations in the country affected by CrowdStrike's outage in July are dropping...
1 hour ago - Experts at the Netherlands Institute for Radio Astronomy (ASTRON) claim that second-generation, or "V2," Mini Starlink satellites emit interference that is a staggering 32 times stronger than that from previous models. Director Jessica...
1 hour ago - The PKfail incident shocked the computer industry, exposing a deeply hidden flaw within the core of modern firmware infrastructure. The researchers who uncovered the issue have returned with new data, offering a more realistic assessment...
1 hour ago - Nighttime anxiety can really mess up your ability to sleep at night. Here's what you can do about it right now.
1 hour ago - With spectacular visuals and incredible combat, I cannot wait for Veilguard to launch on Oct. 31.