Open-Sora/eval
Shen Chenhui d71a06756b Hotfix/vae sampler (#106)
* format

* format

* fix eval loss

* format

* use default seed

* format

* change back ckpt_every to 1k

---------

Co-authored-by: Shen-Chenhui <shen_chenhui@u.nus.edu>
2024-05-21 10:36:26 +08:00
..
human_eval [feat] add lancher for eval loss and search bs 2024-05-15 17:06:54 +00:00
loss Hotfix/vae sampler (#106) 2024-05-21 10:36:26 +08:00
vae
vbench
vbench_i2v
README.md [feat] add lancher for eval loss and search bs 2024-05-15 17:06:54 +00:00
sample.sh

Evalution

Human evaluation

To conduct human evaluation, we need to generate various samples. We provide many prompts in assets/texts, and defined some test setting covering different resolution, duration and aspect ratio in eval/sample.sh. To facilitate the usage of multiple GPUs, we split sampling tasks into several parts.

# image (1)
bash eval/sample.sh /path/to/ckpt -1
# video (2a 2b 2c ...)
bash eval/sample.sh /path/to/ckpt -2a
# launch 8 jobs at once (you must read the script to understand the details)
bash eval/launch.sh /path/to/ckpt

Rectified Flow Loss

Evaluate the rectified flow loss with the following commands.

# image
torchrun --standalone --nproc_per_node 1 eval/loss/eval_loss.py configs/opensora-v1-2/misc/eval_loss.py --data-path /path/to/img.csv --ckpt-path /path/to/ckpt

# video
torchrun --standalone --nproc_per_node 1 eval/loss/eval_loss.py configs/opensora-v1-2/misc/eval_loss.py --data-path /path/to/vid.csv --ckpt-path /path/to/ckpt

# select resolution
torchrun --standalone --nproc_per_node 1 eval/loss/eval_loss.py configs/opensora-v1-2/misc/eval_loss.py --data-path /path/to/vid.csv --ckpt-path /path/to/ckpt --resolution 720p

To launch multiple jobs at once, use the following script.

bash eval/loss/launch.sh /path/to/ckpt

VBench

VBench is a benchmark for short text to video generation. We provide a script for easily generating samples required by VBench.

# vbench tasks (4a 4b 4c ...)
bash eval/sample.sh /path/to/ckpt -4a
# launch 8 jobs at once (you must read the script to understand the details)
bash eval/launch.sh /path/to/ckpt

After generation, install the VBench package according to their instructions. Then, run the following commands to evaluate the generated samples.

bash eval/vbench/vbench.sh /path/to/video_folder

VBench-i2v

VBench-i2v is a benchmark for short image to video generation (beta version).

TBD

VAE

Dependencies

pip install decord
pip install pytorchvideo
pip install lpips
pip install scipy
# Also, if torchvision.transforms.augentation still use `functional_tensor` and cause error,change to use `_functional_tensor`, follow https://blog.csdn.net/lanxing147/article/details/136625264

Commands: carefule to change the setting to training setting

# metric can any one or list of: ssim, psnr, lpips, flolpips
python eval/vae/eval_common_metric.py --batch_size 2 --real_video_dir <path/to/original/videos> --generated_video_dir <path/to/generated/videos> --device cuda --sample_fps 24 --crop_size 256 --resolution 256 --num_frames 17 --sample_rate 1 --metric ssim psnr lpips flolpips