-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathtrain_diffusion.py
More file actions
158 lines (121 loc) · 7.18 KB
/
train_diffusion.py
File metadata and controls
158 lines (121 loc) · 7.18 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
import os
import socket
import torch
import hydra
from omegaconf import DictConfig, OmegaConf
OmegaConf.register_new_resolver("eval", eval)
from ignite.contrib.handlers import ProgressBar
from ignite.engine import Events, Engine
from ignite.metrics import RunningAverage
from ignite.contrib.handlers.tensorboard_logger import TensorboardLogger
from ignite.handlers import Checkpoint, DiskSaver
from src.utils.reproducibility import set_seed
from src.utils.tensorboard import setup_tb_logging_valid, setup_tb_logging_train
from src.train_utils import setup_validation_diffusion, create_train_dataloaders, resume_training, create_model
from src.core.trainer import TrainerDiffusion, setup_scheduler_step
from src.core.diffusion_manager import DiffusionManager
from src.utils.config import load_and_merge_autoenc_cfg, flat_hydra_config, config_exp_folder
from src.data import create_skeleton
from src.utils.load import load_autoencoder
os.environ["CUBLAS_WORKSPACE_CONFIG"] = ":4096:8"
def train(output_log_path: str,
num_epochs: int,
eval_frequency: int = None,
num_iteration_eval: int = 0,
device: str = 'cpu',
seed: int = 63485,
num_iter_perepoch: int = None,
detect_anomaly: bool = False,
if_resume_training: bool = False,
if_run_validation: bool = False,
**kwargs) -> None:
torch.autograd.set_detect_anomaly(detect_anomaly)
# Init seed
set_seed(seed)
torch.set_default_dtype(torch.float64 if kwargs["dtype"]== "float64" else torch.float32)
print(f"Default datatype: {torch.get_default_dtype()} . Using device: {device} - {torch.cuda.get_device_name(0).replace(' ', '_') if torch.cuda.is_available() else ''}")
skeleton = create_skeleton(**kwargs)
# # Create model
model = create_model(device, skeleton=skeleton, num_nodes=skeleton.num_nodes,
node_types=skeleton.nodes_type_id, **kwargs['autoenc_arch'], **kwargs)
print(f"Loaded Autoencoder Model with {str(round(sum(p.numel() for p in model.parameters() if p.requires_grad)/1000000, 3))}M parameters.")
# Load GraphModel
load_autoencoder(model, **kwargs)
diffusionmanager = DiffusionManager(skeleton=skeleton, num_nodes=skeleton.num_nodes,
node_types=skeleton.nodes_type_id, **kwargs)
diffusion = diffusionmanager.get_diffusion().to(device)
print(f"Created Diffusion Model with {str(round(sum(p.numel() for p in diffusion.parameters() if p.requires_grad)/1000000, 3))}M trainable parameters.")
# Load dataset
data_loader_train, data_loader_eval, data_loader_train_eval, random_state_manager = create_train_dataloaders(skeleton=skeleton, if_run_validation=if_run_validation, if_resume_training=if_resume_training, **kwargs)
diffusion_trainer = TrainerDiffusion( diffusion, device=device,
ema_decay = 0.995, # exponential moving average decay
skeleton=skeleton,
autoencoder=model,
**kwargs,
)
def set_epoch_seed(engine: Engine):
set_seed(seed + engine.state.epoch)
random_state_manager.reseed_generator(seed + engine.state.epoch)
def preprocess(engine: Engine):
engine.state.batch = [t.to(device) for t in engine.state.batch[:2]]
# Define Training Engines
trainer = Engine(diffusion_trainer.train_step)
RunningAverage(output_transform=lambda x: x).attach(trainer, 'Loss')
trainer.add_event_handler(Events.ITERATION_STARTED, preprocess)
trainer.add_event_handler(Events.EPOCH_STARTED, set_epoch_seed)
# Set up scheduler
setup_scheduler_step(diffusion_trainer, train_engine=trainer)
# Setup progressbar for training
pbar = ProgressBar()
pbar.attach(trainer, ['Loss'])
# Setup tensorboard logging
tb_logger = TensorboardLogger(log_dir=os.path.join(output_log_path, 'tb'))
setup_tb_logging_train(tb_logger, trainer=trainer, optimizer=diffusion_trainer.opt,
model=diffusion, train_out_transform=lambda out: {"loss": out})
if if_run_validation:
evaluator, evaluator_train = setup_validation_diffusion(model, preprocess, diffusion_trainer, skeleton)
pbar.attach(evaluator)
pbar.attach(evaluator_train)
setup_tb_logging_valid(tb_logger, trainer=trainer, evaluator=evaluator, evaluator_train=evaluator_train,
loss_names=["Loss"], metric_names=['ADE', 'APD', "MPJPE_AVG", "MPJPE", 'LLVar', 'LearningRate'])
# Define objecs to save in checkpoint
objects_to_checkpoint = {**diffusion_trainer.get_checkpoint_object(), **{'trainer': trainer, "random_states": random_state_manager}}
checkpoint_handler = Checkpoint(objects_to_checkpoint, DiskSaver(os.path.join(output_log_path, "checkpoints"), create_dir=True, require_empty=False), n_saved=10,
global_step_transform= lambda engine, event_name: trainer.state.epoch, score_function=lambda engine: -1* round(engine.state.metrics['ADE'], 7), score_name="val_ade")
latest_checkpoint_handler = Checkpoint(objects_to_checkpoint, checkpoint_handler.save_handler, n_saved=1, global_step_transform= lambda engine, event_name: trainer.state.epoch)
if if_resume_training:
checkpoint = torch.load(kwargs['load_path'])
Checkpoint.load_objects(to_load=objects_to_checkpoint, checkpoint=checkpoint)
print("Resuming training from: epoch = ", trainer.state.epoch, '. Iteration = ', trainer.state.iteration)
@trainer.on(Events.EPOCH_COMPLETED)
def save_latest_model(engine: Engine):
latest_checkpoint_handler(trainer)
#Setup evaluation process between epochs
if if_run_validation:
@trainer.on(Events.EPOCH_COMPLETED(every=eval_frequency))
def train_epoch_completed(engine: Engine):
set_seed(0)
random_state_manager.reseed_generator(0)
evaluator.run(data_loader_eval)
evaluator_train.run(data_loader_train_eval, epoch_length=num_iteration_eval)
checkpoint_handler(evaluator)
set_seed(seed + trainer.state.epoch)
random_state_manager.reseed_generator(seed + trainer.state.epoch)
trainer.run(data_loader_train, epoch_length=num_iter_perepoch if num_iter_perepoch is not None else len(data_loader_train), max_epochs=num_epochs)
tb_logger.close()
@hydra.main(config_path="./configs/config_train_diffusion", config_name="config", version_base="1.3")
def main(cfg: DictConfig):
cfg = flat_hydra_config(cfg)
assert cfg['num_workers'] >=1, "Necessary for reproducibility issues"
if cfg['if_resume_training']:
print("Configuration file will be ignored. Considering old configuration file.")
cfg = resume_training(cfg)
else:
# Start from scratch
cfg = config_exp_folder(cfg)
cfg = load_and_merge_autoenc_cfg(cfg)
print("Traing data saved at ", cfg['output_log_path'])
print("Training on device: ", f"{socket.gethostname().split('.')[0]}")
train(**cfg)
if __name__ == '__main__':
main()