Skip to content

An Application to Generate Music and Artwork according to user input

Notifications You must be signed in to change notification settings

Harshath143/AIGEN

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 

Repository files navigation

AIGEN

This project aims to develop an application that generates images and music based on user input using pre-trained deep learning models through an Application Programming Interface (API). The application will allow users to input text and the pre-trained deep learning models will generate a corresponding image and music track.

WORKING ⚔️

The system is divided into two parts: Getting the input from the user, Display the generated output to the user.

The music and image is generated using publicly available pre-trained models through Inference API from HuggingFace.

Image Generating Model: Stable Diffusion is a latent text-to-image diffusion model capable of generating photo-realistic images given any text input

Music Generating Model: Riffusion is a latent text-to-image diffusion model capable of generating spectrogram images given any text input. These spectrograms can be converted into audio clips. image

OUTPUT 👍

APPLICATION HOMEPAGE

image

APPLICATION DISPLAY PAGE

SAMPLE 1 : Input ‘Boat with sunrise’ and ‘Pleasant heavy metal’

image

SAMPLE 2 : Input ‘Sparrow in tree abstract’ and ‘fun disco’

image

TO RUN LOCALLY 👇

  1. Download the repository or clone it locally
  2. Create a python virtual environment using python -m venv
  3. Once Created Activate the virtual environment
  4. Install the packages from the requierments.txt
  5. Run the application using flask --app aigen run

About

An Application to Generate Music and Artwork according to user input

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published