mirror of
https://github.com/hpcaitech/Open-Sora.git
synced 2026-04-10 12:49:38 +02:00
[stage2] update config
This commit is contained in:
parent
5f9e642278
commit
ce601066c5
|
|
@ -88,5 +88,4 @@ grad_clip = 1.0
|
|||
lr = 1e-4
|
||||
ema_decay = 0.99
|
||||
adam_eps = 1e-15
|
||||
warmup_steps = 1000
|
||||
|
||||
|
|
|
|||
|
|
@ -70,6 +70,7 @@ def parse_args(training=False):
|
|||
parser.add_argument("--wandb", default=None, type=bool, help="enable wandb")
|
||||
parser.add_argument("--load", default=None, type=str, help="path to continue training")
|
||||
parser.add_argument("--start-from-scratch", action="store_true", help="start training from scratch")
|
||||
parser.add_argument("--warmup-steps", default=None, type=int, help="warmup steps")
|
||||
|
||||
return parser.parse_args()
|
||||
|
||||
|
|
|
|||
|
|
@ -330,18 +330,18 @@ def main():
|
|||
wandb.log(
|
||||
{
|
||||
"iter": global_step,
|
||||
"acc_step": acc_step,
|
||||
"epoch": epoch,
|
||||
"loss": loss.item(),
|
||||
"avg_loss": avg_loss,
|
||||
"acc_step": acc_step,
|
||||
"lr": optimizer.param_groups[0]["lr"],
|
||||
"move_data_time": move_data_t.elapsed_time,
|
||||
"encode_time": encode_t.elapsed_time,
|
||||
"mask_time": mask_t.elapsed_time,
|
||||
"diffusion_time": loss_t.elapsed_time,
|
||||
"backward_time": backward_t.elapsed_time,
|
||||
"update_ema_time": ema_t.elapsed_time,
|
||||
"reduce_loss_time": reduce_loss_t.elapsed_time,
|
||||
"debug/move_data_time": move_data_t.elapsed_time,
|
||||
"debug/encode_time": encode_t.elapsed_time,
|
||||
"debug/mask_time": mask_t.elapsed_time,
|
||||
"debug/diffusion_time": loss_t.elapsed_time,
|
||||
"debug/backward_time": backward_t.elapsed_time,
|
||||
"debug/update_ema_time": ema_t.elapsed_time,
|
||||
"debug/reduce_loss_time": reduce_loss_t.elapsed_time,
|
||||
},
|
||||
step=global_step,
|
||||
)
|
||||
|
|
|
|||
Loading…
Reference in a new issue