-
Notifications
You must be signed in to change notification settings - Fork 24
/
airformer_trainer.py
40 lines (36 loc) · 1.67 KB
/
airformer_trainer.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
import numpy as np
import torch
from torch.optim.lr_scheduler import MultiStepLR
from torch.optim import Adam
from src.base.trainer import BaseTrainer
from src.utils import graph_algo
class Trainer(BaseTrainer):
def __init__(self, **args):
super(Trainer, self).__init__(**args)
self._optimizer = Adam(self.model.parameters(), self._base_lr)
self._supports = self._calculate_supports(args['adj_mat'], args['filter_type'])
self._lr_scheduler = MultiStepLR(self._optimizer,
self._steps,
gamma=self._lr_decay_ratio)
def _calculate_supports(self, adj_mat, filter_type):
num_nodes = adj_mat.shape[0]
new_adj = adj_mat + np.eye(num_nodes)
if filter_type == "scalap":
supports =[graph_algo.calculate_scaled_laplacian(new_adj).todense()]
elif filter_type == "normlap":
supports =[graph_algo.calculate_normalized_laplacian(
new_adj).astype(np.float32).todense()]
elif filter_type == "symnadj":
supports =[graph_algo.sym_adj(new_adj)]
elif filter_type == "transition":
supports =[graph_algo.asym_adj(new_adj)]
elif filter_type == "doubletransition":
supports =[graph_algo.asym_adj(new_adj),
graph_algo.asym_adj(np.transpose(new_adj))]
elif filter_type == "identity":
supports =[np.diag(np.ones(new_adj.shape[0])).astype(np.float32)]
else:
error = 0
assert error, "adj type not defined"
supports = [torch.tensor(i).cuda() for i in supports]
return supports