ECCV 2026 Official Workshop

Efficient Visual Generation

A workshop on efficient generative models for visual and multimodal generation, spanning lightweight architectures, post-training methods, scalable deployment, real-time inference, long video systems, and world model applications.

Overview

In recent years, visual generation has advanced rapidly and reshaped research at major computer vision conferences such as ECCV, ICCV, and CVPR. While generative models can now produce compelling images and videos, making them efficient remains a key bottleneck as compute, latency, and energy constraints increasingly matter.

Efficient visual generative models are essential for scalable, real-time applications across image and video generation, XR, gaming, design, and creative tooling. In practice, these systems must run across diverse environments, ranging from edge devices and mobile platforms to cloud services and interactive production pipelines.

The urgency has become even more pronounced with the rise of video generation and world models. Compared with image synthesis, long-horizon video generation places far greater pressure on compute, memory, and bandwidth, while world models introduce additional demands from real-time interaction, long-context reasoning, and repeated rollouts.

As generative systems move from research prototypes to product-facing deployments, efficiency is no longer just an optimization target. It becomes a core enabler that determines feasibility, responsiveness, operating cost, and energy usage in latency-sensitive and resource-constrained settings.

EVG highlights architectural innovation, data-efficient learning, resource-aware inference, and deployment-ready optimization strategies that can accelerate the adoption of generative models in real-world applications without compromising output quality and controllability.

Workshop Themes

What the workshop will cover

01

Efficient Model Architectures

Architectures that reduce compute and memory without sacrificing generative quality.

  • Lightweight backbones for image, video, and multimodal generation
  • Token, latent, and memory compression strategies
  • Adaptive computation and resource-aware model design

02

Efficient Training and Post-Training

Methods that reduce the cost of adaptation, scaling, and refinement for generative systems.

  • Data-efficient training, finetuning, and adaptation
  • Post-training optimization for controllability and quality
  • Distillation and alignment with limited compute budgets

03

Inference and Deployment

Practical methods for low-latency inference and deployment across edge and cloud platforms.

  • Quantization, pruning, and model compression for generation
  • Serving systems for scalable and real-time inference
  • Deployment strategies for mobile, edge, and cloud settings

04

Application-Specific Efficiency

Efficiency challenges that emerge in demanding real-world generative applications.

  • Streaming and interactive generation pipelines
  • Long video, high-resolution, and temporally consistent synthesis
  • Efficient world models and embodied rollout systems

Invited Speakers

Confirmed speakers

Portrait of Gal Chechik
Confirmed Speaker

Gal Chechik

Professor, Bar-Ilan University; Sr. Director, NVIDIA Research, Israel

Portrait of Xun Huang
Confirmed Speaker

Xun Huang

Research Scientist and entrepreneur; formerly Adobe Research, CMU, and NVIDIA

Organizers

Portrait of Kai Wang

Kai Wang

Assistant Professor, City University of Hong Kong, Dongguan, China

Contact

Inquiries and collaboration

For questions about invited talks, submissions, or future updates, please contact the organizing team through the primary contact below.

Email Shiqi Yang