mirror of
https://github.com/hpcaitech/Open-Sora.git
synced 2026-04-10 04:37:45 +02:00
[docs] update
This commit is contained in:
parent
a887f54711
commit
9117c9fcc7
|
|
@ -75,6 +75,8 @@ Open-Sora 1.2 starts from the [PixArt-Σ 2K](https://github.com/PixArt-alpha/Pix
|
|||
|
||||
After the above adaptation, we are ready to train the model on videos. The adaptation above maintains the original model's ability to generate high-quality images.
|
||||
|
||||
With rectified flow, we can reduce the number of sampling steps for video from 100 to 30, which greatly reduces the waiting time for inference.
|
||||
|
||||
## More data and better multi-stage training
|
||||
|
||||
Due to a limited computational budget, we carefully arrange the training data from low to high quality and split our training into three stages. Our training involves 12x8 GPUs, and the total training time is about 2 weeks.
|
||||
|
|
@ -130,3 +132,5 @@ We sampled 1k videos from pixabay as validation dataset. We calculate the evalua
|
|||
In addition, we also keep track of [VBench](https://vchitect.github.io/VBench-project/) scores during training. VBench is an automatic video evaluation benchmark for short video generation. We calcuate the vbench score with 240p 2s videos. The two metrics verify that our model continues to improve during training.
|
||||
|
||||

|
||||
|
||||
All the evaluation code is released in `eval` folder. Check the [README](/eval/README.md) for more details.
|
||||
|
|
|
|||
|
|
@ -476,10 +476,12 @@ NUM_FRAMES_MAP = {
|
|||
"2x": 102,
|
||||
"4x": 204,
|
||||
"8x": 408,
|
||||
"16x": 816,
|
||||
"2s": 51,
|
||||
"4s": 102,
|
||||
"8s": 204,
|
||||
"16s": 408,
|
||||
"32s": 816,
|
||||
}
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Reference in a new issue