Skip to content

Commit 862def1

Browse files
committed
update inspiremusic
1 parent f84ec4d commit 862def1

File tree

1 file changed

+14
-17
lines changed

1 file changed

+14
-17
lines changed

inspiremusic/index.html

Lines changed: 14 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626

2727
<div class="container pt-5 mt-5 shadow-lg p-5 mb-5 bg-white rounded">
2828
<div class="text-center">
29-
<h2>InspireMusic: A Unified Framework for Controlled High-Fidelity Long-Form Music, Song and Audio Generation with Dual Audio Tokenizations</h2>
29+
<h2>InspireMusic: A Unified Framework for Controlled High-Fidelity Long-Form Music, Song and Audio Generation</h2>
3030

3131
<!-- [<a href="https://arxiv.org/abs/2407.04051">Paper</a>]-->
3232
[<a href="https://github.com/FunAudioLLM/InspireMusic">Code</a>]
@@ -44,13 +44,7 @@ <h2>InspireMusic: A Unified Framework for Controlled High-Fidelity Long-Form Mus
4444
</div>
4545
<p><b>Abstract</b>
4646

47-
This report introduces <b>InspireMusic</b>, a unified framework for music generation that combines semantic and acoustic tokens with an autoregressive
48-
Transformer and conditional flow-matching modeling to create expressive high-quality audio.
49-
InspireMusic utilizes an audio tokenizer designed to produce discrete tokens with a single codebook that captures both acoustic and semantic characteristics of the audio waveform.
50-
To maintain musical structure, inputs include musical form and timestamps. These tokens are processed by an audio language model configured as an autoregressive transformer,
51-
to learn and predict intricate musical patterns. Subsequently, the generated tokens are refined using a flow-matching model with
52-
a sophisticated acoustic neural codec model with multiple codebooks, yielding high-fidelity audio.
53-
This approach ensures the generated music is coherent and rich, facilitating tasks in music, song, and audio generation.</p>
47+
Recent advances in generative modeling have transformed the landscape of music and audio generation. In this work, we introduce <b>InspireMusic</b>, a unified framework designed to generate high-fidelity music, songs, and audio, which integrates an autoregressive transformer with a super-resolution flow-matching model. This framework enables the direct generation of high-fidelity long-form audio at 48kHz from both text and audio modalities. Unlike prior systems that focus solely on symbolic or raw audio generation, our approach employs dual audio tokenizers to capture both the global musical structure and the fine-grained acoustic details, allowing for high quality audio generation with long-form coherence. This framework represents a significant advancement in music generation by directly modeling raw audio, ensuring both diversity and high-fidelity output.</p>
5448
</p>
5549

5650
<p><b>Highlights</b>
@@ -101,15 +95,18 @@ <h2 id="InspireMusic-overview" style="text-align: center;">Overview of InspireMu
10195
</p>
10296
</body>
10397
<p style="text-align: center;" >
104-
<b>Figure 1.</b> An overview of the InspireMusic framework.
105-
We introduce InspireMusic, a unified framework for music, song and audio generation, capable of producing 48kHz long-form audio.
106-
InspireMusic employs an autoregressive transformer to generate music tokens in response to textual input. Complementing this, an ODE-based diffusion model, specifically flow matching, is utilized to reconstruct latent features from these generated music tokens.
107-
Then a vocoder generates audio waveforms from the reconstructed features.
108-
for input text, an ODE-based diffusion model, flow matching,
109-
to reconstruct latent features from the generated music tokens,
110-
and a vocoder to generate audio waveforms. InspireMusic is capable of text-to-music, music continuation, music reconstruction, and music super resolution tasks.
111-
It employs WavTokenizer as an audio tokenizer to convert 24kHz audio into 75Hz discrete tokens, while HifiCodec serves as a music tokenizer, transforming 48kHz audio into 150Hz latent features compatible with the flow matching model.
112-
<!-- [<a href="https://arxiv.org/abs/">Paper</a>]-->
98+
<b>Figure 1.</b> An overview of the InspireMusic framework. We introduce InspireMusic, a unified framework for music, song, and audio generation capable of producing high-quality 48kHz long-form audio. InspireMusic consists of three key components:
99+
100+
- **Dual Audio Tokenizers**:
101+
The framework first converts raw audio waveforms into discrete tokens that are efficiently processed by the autoregressive model. We employ two tokenizers: WavTokenizer converts 24kHz audio into 75Hz discrete tokens, while Hifi-Codec transforms 48kHz audio into 150Hz latent features suited for our flow matching model.
102+
103+
- **Autoregressive Transformer**:
104+
This component is trained using a next-token prediction approach on both text and audio tokens, enabling it to generate coherent and contextually relevant audio sequences.
105+
106+
- **Super-Resolution Flow Matching** Model:
107+
An ODE-based diffusion model, specifically a super-resolution flow matching (SRFM) model, maps the lower-resolution audio tokens to latent features with a higher sampling rate. A vocoder then generates the final audio waveform from these enhanced latent features.
108+
109+
InspireMusic supports a range of tasks including text-to-music, music continuation, music reconstruction, and music super-resolution.-- [<a href="https://arxiv.org/abs/">Paper</a>]-->
113110
</p>
114111
</div>
115112

0 commit comments

Comments
 (0)