Feature request / 功能建议
怎么设置可以在推理的时候模型并行呢?
How can I set up model parallelism during inference?
Motivation / 动机
80G,视频总帧数320
args:
mode: inference
i2v_encode_video: True
latent_channels: 16
load: SCAIL-Preview/model
batch_size: 1
input_type: txt
input_file: input.txt
sampling_image_size: [512, 896]
vae_compress_size: [4, 8, 8] # t h w
bf16: True
sampling_fps: 32
force_inference: True
sequence_parallel_size: 1
image2video: True
use_i2v_clip: True
use_pose: True
representation: 'smpl_downsample'
save_tag: official
output_dir: SCAIL/examples
爆显存了。
Your contribution / 您的贡献
。