Skip to content

vllm-project/vLLM-in-PyTorch-Conference-2025

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 

Repository files navigation

vLLM-in-PyTorch-Conference-2025

There are at least 53 videos (out of the full 117 videos in PyTorch conference 2025) mentioning vLLM. The ratio is 45%, or around 50%. The list is non-exhaustive. If you find any other videos mentioning vLLM in PyTorch conference 2025, welcome to submit a pull-request to add it in the list!

  1. Keynote: Welcome & Opening Remarks Matt White, Executive Director, PyTorch Foundation
  1. Keynote: vLLM & Deepspeed Updates Simon Mo & Tunji Ruwase
  1. Keynote: Ray: A Distributed Compute Engine for AI Robert Nishihara & Ion Stoica
  1. Keynote: Olmo-Thinking: Training a Fully Open Reasoning Model Nathan Lambert
  1. Open Source Model Performance Optimization With SGLang Yineng Zhang, Together AI
  1. PyTorch Symmetric Memory: A New Programming Paradigm for Distributed AI Ke Wen & Chien-Chin Huang
  1. Sponsored Session: Lightning Talk: Accelerating Experimentation and Unlocking Real-Time Inference on Microcontrollers with Lightning Niall Lyons & Luca Antiga
  1. Deploying GenAI for Audio Generation on Mobile CPUs With ExecuTorch Gian Marco Iodice, Arm
  1. Sponsored Session: Beyond the Node: Scaling Inference with Cluster-Wide KVCache Management Alon Yariv
  1. Lightning Talk: Hardware-Aware Python Packages ~ PyTorch and WheelNext Grab the Wheel! Jonathan Dekhtiar & Eli Uriegas
  1. Our Journey With TorchTitan Linsong Chu & Garrett Goon, IBM Research
  1. The Building Blocks of Agentic Al Joe Spisak, Product Director, Meta Superintelligence Labs
  1. Sponsor Keynote: Build AI Anywhere with ROCm™ Software: AMD and PyTorch Bring Cloud-to-Client Power to Developers Anush Elangovan
  1. Keynote: PyTorch Technical Deep Dive Alban Desmaison, Peng Wu, Mark Saroufim & Edward Yang, Meta
  1. Sponsored Session: Lightning Talk: Build and Deploy AI Flows with an Agent Factory Arjun Upadhyay
  1. Sponsored Session: Everything Everywhere all at Once: vLLM Hardware Optionality with Spotify and Google Brittany Rockwell & Shireen Kheradpey
  1. An Open Source Post-Training Stack: Kubernetes + Ray + PyTorch + vLLM Robert Nishihara, Anyscale
  1. PyTorch-Native Stack for Agents Allen Wang & Davide Testuggine, Meta
  1. Sponsored Session: Lightning Talk: Optimizing Model Inference with PyTorch 2.0 Devansh Ghatak
  1. Thunder: Distribute and Optimize Your PyTorch Models With Zero Code Changes Luca Antiga & Thomas Viehmann
  1. vLLM: Easy, Fast, and Cheap LLM Serving for Everyone Simon Mo, vLLM
  1. Sponsored Session: Amazingly Fast and Incredibly Scalable Inference with NVIDIA's Dynamo and TensorRT-LLM Harry Kim & Laikh Tewari
  1. Lightning Talk: Unlock the Future of Generative AI: TorchTitan's Latest Breakthroughs Tianyu Liu & Jiani Wang
  1. Breaking Heterogeneity Barriers: Unified Cloud-to-Robot AI System SW Stack for Embodied Intelligence Yonghua Lin
  1. Lightning Talk: Flex Attention for Inference Boyuan Feng & Driss Guessous, Meta
  1. Sponsored Session: Lightning Talk: Improving Drug Discovery with Machine Learning and Molecular Dynamics Simon Axelrod
  1. Sponsored Session: Lightning Talk: Accelerated Software for a Post-Moore World Jay Dawani
  1. Sponsored Session: Building the Next Generation of Open Source AI Tooling Travis Oliphant
  1. No GPU Left Behind: Scaling Online LLM Training With Co-located VLLM in TRL Mert Toslali & Yu Chin Fabian Lim
  1. Lightning Talk: Challenges and Standardization in PyTorch Ecosystem Accelerators Zesheng Zong & Ashok Emani
  1. Keynote: Building the Open Agent Ecosystem Together: Introducing OpenEnvJoe Spisak, Lysandre Debut
  1. Enabling VLLM V1 on AMD GPUs With Triton Thomas Parnell, IBM Research & Aleksandr Malyshev, AMD
  1. Transformers: Standardizing Model Definitions Across the PyTorch Ecosystem L. Debut & A. Zucker
  1. How Modern PyTorch Supercharges Multimodal Training and Inference at Luma AI Thomas Neff, Luma AI
  1. Optimizing Long-Tail and MoE Challenges in Reinforcement Learning with SGLang Chenyang Zhao, UCLA
  1. Serving PyTorch LLMs at Scale: Disaggregated Inference With Kubernetes and Llm-d M. Ayoub & C. Liu
  1. Verl: A Flexible and Efficient RL Framework for LLMs Hongpeng Guo & Ziheng Jiang, ByteDance Seed
  1. Sponsored Session: Lightning Talk: PyTorch and Democratization of AI Accelerators Hong-Seok Kim
  1. Multi-Accelerator PyTorch Serving With NxD Inference and vLLM Yahav Biran & Liangfu Chen, Amazon
  1. Scaling KV Caches for LLMs: How LMCache + NIXL Handle Network and Storage Heterogeneity J. Jiang & M. Khazraee
  1. Sponsored Session: Lightning Talk: PyTorch in Production: Boosting LLM Training and Inferencing on Ascend NPU F. Hua
  1. Keynote Panel: Hardwares & Accelerators D Patel, S Zhou, P Salanki, N Perumbeti, M Saroufim
  1. Blazing Fast GenAI Inference With Torch.compile Richard Zou, Meta
  1. Sponsored Session: Empowering AI Everywhere: Democratizing PyTorch with Intel AIPCs & More F. Zhao & E. Wang
  1. Designing and Building Custom Reinforcement Learning Environments for Fine-tuning LLMs N. Bantilan
  1. Sponsored Session: Accelerating GenAI Inference: From AWS Deep Learning Containers to Scaling Amazon Rufus on Trainium P. Nguyen & A. Zhao
  1. Lightning Talk: Vllm-triton-backend: How To Get State-of-the-art Performance on NVIDIA and AMD With Just Triton B. Ringlein
  1. Sponsored Session: Lightning Talk: Efficient Inference Serving with Kubernetes Gateway API Inference Extension Lin Sun
  1. Lightning Talk: Improved GEMM and SDPA Performance on ROCm With Composable Kernel Andres Lugo, AMD
  1. Kimi K2 and Our Contributions to Open Source Yuxin Wu, Moonshot AI
  1. Lightning Talk: A Stable Limited LibTorch ABI? How?! (and Why?) Jane Xu, Meta
  1. Sponsored Session: Arctic Inference: Breaking the Speed-Cost Tradeoff in LLM Serving Aurick Qiao
  1. PyTorch Conference 2025 Recap

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published