Magenta RealTime: An Open-Weights Live Music Model (via) Fun new "live music model" release from Google DeepMind:
Today, we’re happy to share a research preview of Magenta RealTime (Magenta RT), an open-weights live music model that allows you to interactively create, control and perform music in the moment. [...]
As an open-weights model, Magenta RT is targeted towards eventually running locally on consumer hardware (currently runs on free-tier Colab TPUs). It is an 800 million parameter autoregressive transformer model trained on ~190k hours of stock music from multiple sources, mostly instrumental.
No details on what that training data was yet, hopefully they'll describe that in the forthcoming paper.
For the moment the code is on GitHub under an Apache 2.0 license and the weights are on HuggingFace under Creative Commons Attribution 4.0 International.
The easiest way to try the model out is using the provided Magenta_RT_Demo.ipynb Colab notebook.
It takes about ten minutes to set up, but once you've run the first few cells you can start interacting with it like this:
Recent articles
- Highlights from my appearance on the Data Renegades podcast with CL Kao and Dori Wilson - 26th November 2025
- Claude Opus 4.5, and why evaluating new LLMs is increasingly difficult - 24th November 2025
- sqlite-utils 4.0a1 has several (minor) backwards incompatible changes - 24th November 2025