Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Riffusion

From Wikipedia, the free encyclopedia
Music-generating machine learning model
Riffusion
Developers
  • Seth Forsgren
  • Hayk Martiros
Initial releaseDecember 15, 2022
Repositorygithub.com/hmartiro/riffusion-inference
Written inPython
TypeText-to-image model
LicenseMIT License
Websiteriffusion.com
Generated spectrogram from the prompt "bossa nova withelectric guitar" (top), and the resulting audio after conversion (bottom)

Riffusion is aneural network, designed by Seth Forsgren and Hayk Martiros, that generates music using images of sound rather than audio.[1]

The resulting music has been described as "de otro mundo" (otherworldly),[2] although unlikely to replace man-made music.[2] The model was made available on December 15, 2022, with the code also freely available onGitHub.[3]

The first version of Riffusion was created as afine-tuning ofStable Diffusion, an existing open-source model for generating images from text prompts, onspectrograms,[1] resulting in a model which used text prompts to generate image files which could then be put through aninverse Fourier transform and converted into audio files.[3] While these files were only several seconds long, the model could also uselatent space between outputs tointerpolate different files together[1][4] (using theimg2img capabilities of SD).[5] It was one of many models derived from Stable Diffusion.[5]

In December 2022, Mubert[6] similarly used Stable Diffusion to turn descriptive text into music loops. In January 2023, Google published a paper on their own text-to-music generator called MusicLM.[7][8]

Forsgren and Martiros formed a startup, also called Riffusion, and raised $4 million in venture capital funding in October 2023.[9][10]

References

[edit]
  1. ^abcColdewey, Devin (December 15, 2022)."Try 'Riffusion,' an AI model that composes music by visualizing it".
  2. ^abLlano, Eutropio (December 15, 2022)."El generador de imágenes AI también puede producir música (con resultados de otro mundo)".
  3. ^abNasi, Michele (December 15, 2022)."Riffusion: creare tracce audio con l'intelligenza artificiale".IlSoftware.it.
  4. ^"Essayez "Riffusion", un modèle d'IA qui compose de la musique en la visualisant". December 15, 2022.
  5. ^ab"文章に沿った楽曲を自動生成してくれるAI「Riffusion」登場、画像生成AI「Stable Diffusion」ベースで誰でも自由に利用可能".GIGAZINE. 16 December 2022.
  6. ^"Mubert launches Text-to-Music interface – a completely new way to generate music from a single text prompt". December 21, 2022.
  7. ^"MusicLM: Generating Music From Text". January 26, 2023.
  8. ^"5 Reasons Google's MusicLM AI Text-to-Music App is Different". January 27, 2023.
  9. ^Gal, Dr. Itay (February 10, 2025)."Free A.I. music creation platform launches, competing with Suno".The Jerusalem Post. RetrievedFebruary 16, 2025.
  10. ^Nuñez, Michael (January 30, 2025)."Riffusion's free AI music platform could be the Spotify of the future".VentureBeat. RetrievedFebruary 16, 2025.
Free
Subscription
Hybrid
Gen AI
Defunct
Concepts
Chatbots
Models
Text
Coding
Image
Video
Speech
Music
Controversies
Agents
Companies
Concepts
Applications
Implementations
Audio–visual
Text
Decisional
People
Architectures


Stub icon

Thisartificial neural network-related article is astub. You can help Wikipedia byexpanding it.

Stub icon

Thisscientific software article is astub. You can help Wikipedia byexpanding it.

Retrieved from "https://en.wikipedia.org/w/index.php?title=Riffusion&oldid=1323145739"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp