Skip to content

longxiang-ai/awesome-video-diffusions

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

24 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

Awesome Video Diffusions Awesome

A curated list of latest research papers, projects and resources related to Video Diffusion Models and Video Generation. Content is automatically updated daily.

Last Update: 2026-03-05 02:07:49

πŸ“° Latest Updates

πŸš€ [2026-02] Project Launched β€” v1.0

  • Adapted from awesome-gaussians framework for tracking video diffusion research

  • Unified CLI: Single entry point python main.py with subcommands: init, search, suggest, export-bib, readme

  • Interactive Configuration Wizard: Run python main.py init to set up keywords, domains, time range, and API keys step-by-step

  • Custom Time Range Filtering: Support relative periods (6m, 1y, 2y) and absolute date ranges

  • Smart Link Extraction: Automatically extracts and classifies GitHub, project page, dataset, video, demo, and HuggingFace links from paper abstracts

  • BibTeX Export: Fetch BibTeX from arXiv and export to .bib files with category/date filters

  • LLM Keyword Suggestion: Paste a few paper titles or arXiv IDs, and an LLM automatically generates optimized search keywords

  • arXiv Domain Filtering: Restrict searches to specific arXiv categories (e.g., cs.CV, cs.AI, cs.MM)

  • 16 Research Categories: Comprehensive taxonomy covering T2V, I2V, video editing, controllable generation, world models, and more

  • View detailed updates: News.md πŸ“‹


Categories

Table of Contents

Categorized Papers

3D-aware Video Generation

Applications

Showing the latest 50 out of 53 papers

Architecture & Efficiency

Showing the latest 50 out of 362 papers

Audio & Multi-modal

Controllable Generation

Showing the latest 50 out of 113 papers

Human & Character Animation

Image-to-Video Generation

Long Video Generation

Showing the latest 50 out of 128 papers

Personalization & Customization

Showing the latest 50 out of 95 papers

Physical Understanding

Showing the latest 50 out of 124 papers

Surveys & Benchmarks

Showing the latest 50 out of 195 papers

Text-to-Video Generation

Showing the latest 50 out of 75 papers

Video Editing

Video Inpainting & Completion

Video Super-Resolution & Enhancement

Showing the latest 50 out of 57 papers

World Models & Simulation

Showing the latest 50 out of 102 papers

Classic Papers

Open Source Projects

Tutorials & Blogs

πŸ“‹ Project Features

πŸ› οΈ Core Features

  • Unified CLI (main.py): Single entry point with init, search, suggest, export-bib, readme subcommands
  • Interactive Config Wizard: Guided setup for keywords, domains, time range, and API keys via python main.py init
  • Custom Search Keywords: Configure keywords for title, abstract, or both; with arXiv domain filtering (cs.CV, cs.AI, cs.MM, etc.)
  • Time Range Filtering: Relative periods (30d, 6m, 1y, 2y) or absolute date ranges (YYYY-MM-DD to YYYY-MM-DD)
  • Smart Link Extraction: Auto-classifies URLs from abstracts into GitHub, project page, dataset, video, demo, HuggingFace links
  • BibTeX Export: Fetch BibTeX from arXiv official API; export to .bib files with category and date filters
  • LLM Keyword Suggestion: Input paper titles or arXiv IDs to auto-generate optimized search keywords via OpenAI-compatible API
  • Automated Paper Collection: Daily automatic crawling with GitHub Actions
  • Intelligent Classification: Auto-categorize papers into 16 topics (T2V, I2V, Video Editing, Controllable Generation, World Models, etc.)

πŸ› οΈ Technical Features

  • Robust Error Handling: Multi-layer retry and fallback strategies ensure stable operation
  • GitHub Actions Integration: Automated CI/CD workflows for daily updates
  • Multi-type Link Badges: README entries display PDF, GitHub (with stars), Project, Dataset, Video, Demo, HuggingFace, and Citation badges
  • Detailed Logging: Comprehensive logging for debugging and monitoring
  • Cross-Platform: Support for Windows/Linux/macOS

πŸ“š Data Output

  • Paper JSON files (data/papers_YYYY-MM-DD.json): Full paper metadata with title, authors, abstract, links, keywords, BibTeX
  • BibTeX files (output/*.bib): Ready-to-use bibliography files for LaTeX
  • Auto-generated README: Categorized and formatted paper listings

πŸš€ Quick Start

1. Install Dependencies

pip install -r requirements.txt

2. Interactive Setup (Recommended)

python main.py init

This wizard walks you through:

  • Setting search keywords (for title, abstract, or both)
  • Selecting arXiv domains (e.g., cs.CV, cs.AI, cs.MM)
  • Configuring time range (relative like 6m/1y, or absolute dates)
  • Setting max results
  • Optionally configuring an OpenAI-compatible API key for keyword suggestion

3. Search Papers

# Search with settings from user_config.json
python main.py search

# Override: fetch 200 papers from the last 6 months, include BibTeX
python main.py search --max-results 200 --recent 6m --bibtex

# Search with absolute date range
python main.py search --date-from 2024-01-01 --date-to 2025-01-01

# Include citation counts from Semantic Scholar
python main.py search --citations

4. Export BibTeX

# Export all papers from the latest data file
python main.py export-bib --output output/references.bib

# Export only "Text-to-Video Generation" papers
python main.py export-bib --category "Text-to-Video Generation" --output output/t2v.bib

# Export papers from a specific date range
python main.py export-bib --date-from 2024-06-01 --date-to 2025-01-01 --output output/recent.bib

5. LLM Keyword Suggestion

# Generate keywords from paper titles
python main.py suggest --titles "Video Diffusion Models" "Stable Video Diffusion"

# Generate from arXiv IDs (auto-fetches titles)
python main.py suggest --arxiv-ids 2204.03458 2311.15127

# Auto-write suggested keywords to config
python main.py suggest --titles "Sora" "CogVideoX" --apply

# Use a custom API endpoint (e.g., DeepSeek)
python main.py suggest --titles "Paper Title" --base-url https://api.deepseek.com/v1 --api-key sk-xxx --model deepseek-chat

6. Generate README

# Basic README
python main.py readme

# Include latest papers section and abstracts
python main.py readme --show-latest --show-abstracts

Configuration File

All settings are stored in data/user_config.json:

{
  "search": {
    "keywords": {
      "both_abstract_and_title": ["video diffusion", "video generation", "text-to-video"],
      "abstract_only": ["diffusion model video generation"],
      "title_only": ["video generation", "video diffusion"]
    },
    "domains": ["cs.CV", "cs.AI", "cs.MM"],
    "time_range": {
      "mode": "relative",
      "relative": "1y"
    },
    "max_results": 500
  },
  "api_keys": {
    "openai_api_key": "",
    "openai_base_url": "https://api.openai.com/v1",
    "openai_model": "gpt-4o-mini"
  }
}

Contribution Guidelines

Feel free to submit Pull Requests to improve this list! Please follow these formats:

  • Paper entry format: **[Paper Title](link)** - Brief description
  • Project entry format: [Project Name](link) - Project description

License

CC0

About

A curated and auto-updated collection of video diffusion / video generation papers from arXiv, covering text-to-video, image-to-video, controllable generation, world models, video editing, and 16+ research categories.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages