Discrete Audio IsoFLOP Models
Collection
IsoFLOP Models trained on Yodas+Emilia+Nemotron from budgets of 3e18 to 3e20 • 64 items • Updated
A suite of discrete audio models trained for our IsoFLOP study as part of SODA, which is a unified next-token prediction on interleaved semantic, acoustic, and text tokens.
🥤 Project Page: https://soda-audio.github.io
For full usage instructions (e.g., inference code), and more information, please refer to the SODA-4B-base model card.
The details for this particular model is as follows:
compute_budget: 3e20param_count (non-embedding): 1.68Bhidden_dim: 1920num_layers: 19batch_size: 128training_step: 56131hash_key: a41e32📈 WandB: https://wandb.ai/potsawee/marin/groups/IsoFlop/workspace