-
Notifications
You must be signed in to change notification settings - Fork 0
/
test.py
91 lines (74 loc) · 2.61 KB
/
test.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
"""
Testing scripts.
Authors: Hongjie Fang.
"""
import os
import yaml
import torch
import logging
import warnings
import argparse
import numpy as np
import torch.nn as nn
from tqdm import tqdm
from utils.logger import ColoredLogger
from utils.builder import ConfigBuilder
from utils.functions import to_device
from time import perf_counter
logging.setLoggerClass(ColoredLogger)
logger = logging.getLogger(__name__)
warnings.simplefilter("ignore", UserWarning)
parser = argparse.ArgumentParser()
parser.add_argument(
'--cfg', '-c',
default = os.path.join('configs', 'inference.yaml'),
help = 'path to the configuration file',
type = str
)
args = parser.parse_args()
cfg_filename = args.cfg
with open(cfg_filename, 'r') as cfg_file:
cfg_params = yaml.load(cfg_file, Loader = yaml.FullLoader)
builder = ConfigBuilder(**cfg_params)
logger.info('Building models ...')
model = builder.get_model()
device = torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')
model.to(device)
logger.info('Building dataloaders ...')
test_dataloader = builder.get_dataloader(split = 'test')
logger.info('Checking checkpoints ...')
stats_dir = builder.get_stats_dir()
checkpoint_file = os.path.join(stats_dir, 'checkpoint.tar')
if os.path.isfile(checkpoint_file):
checkpoint = torch.load(checkpoint_file, map_location={'cuda:5': 'cuda:0'})
model.load_state_dict(checkpoint['model_state_dict'])
start_epoch = checkpoint['epoch']
logger.info("Checkpoint {} (epoch {}) loaded.".format(checkpoint_file, start_epoch))
else:
raise FileNotFoundError('No checkpoint.')
metrics = builder.get_metrics()
def test():
logger.info('Start testing process.')
model.eval()
metrics.clear()
running_time = []
with tqdm(test_dataloader) as pbar:
for data_dict in pbar:
data_dict = to_device(data_dict, device)
with torch.no_grad():
time_start = perf_counter()
res = model(data_dict['rgb'], data_dict['depth'])
time_end = perf_counter()
n, h, w = data_dict['depth'].shape
data_dict['pred'] = res.view(n, h, w)
_ = metrics.evaluate_batch(data_dict, record = True)
duration = time_end - time_start
pbar.set_description('Time: {:.4f}s'.format(duration))
running_time.append(duration)
avg_running_time = np.stack(running_time).mean()
logger.info('Finish testing process, average running time: {:.4f}s'.format(avg_running_time))
metrics_result = metrics.get_results()
metrics.display_results()
return metrics_result
if __name__ == '__main__':
test()