--- datasets: - teticio/audio-diffusion-256 library_name: diffusers --- # Variational Autoencoder Conditioned Diffusion Model This model is designed to generate music tracks based on input playlists by extracting the "taste" from the playlists using a combination of a Variational Autoencoder (VAE) and a conditioned diffusion model. ## Model Details - **VAE**: Learns a compressed latent space representation of the input data, specifically mel spectrogram images of audio samples. - **Diffusion Model**: Generates new data points by progressively refining random noise into meaningful data, conditioned on the VAE's latent space.