Skip to content

Text2Motion is a project that generates videos from textual prompts using diffusion models. This project leverages advanced model offloading and VAE slicing techniques for efficient video synthesis, allowing users to create custom animations directly from text descriptions."

Notifications You must be signed in to change notification settings

iamprabhanjan/Text2Motion

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Text-to-Video Diffusion Model

This project implements a text-to-video diffusion model using the Hugging Face Diffusers library. The model generates short videos based on textual prompts. This project leverages pre-trained diffusion models and fine-tunes them to create high-quality video content.

Features

  • Text-to-Video Generation: Generate videos from text prompts using a diffusion model.
  • Model Offloading: Efficiently manages memory by offloading models to CPU and slicing the VAE.
  • Customizable Output: Allows users to specify the number of frames and apply negative prompts for fine-tuning quality.

Project Structure

  • notebook/: Contains the Colab notebook used for generating the video.
  • src/: Source code files (if any additional scripts are included).
  • requirements.txt: Lists all Python dependencies needed to run the project.
  • README.md: This documentation file.

Installation

To install the required dependencies, run the following:

pip install -r requirements.txt

About

Text2Motion is a project that generates videos from textual prompts using diffusion models. This project leverages advanced model offloading and VAE slicing techniques for efficient video synthesis, allowing users to create custom animations directly from text descriptions."

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published