You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
D:\OneDrive - Deakin University\Desktop\visual_token_matching>python main.py --stage 0 --task_fold 1
C:\Users\rajar\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\amp\autocast_mode.py:204: UserWarning: User provided device_type of 'cuda', but CUDA is not available. Disabling
warnings.warn('User provided device_type of 'cuda', but CUDA is not available. Disabling')
Load ckpt from model/pretrained_checkpoints\beit_base_patch16_224_pt22k.pth
Load state_dict by model_key = model
Expand the shared relative position embedding to each transformer block.
Traceback (most recent call last):
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\main.py", line 75, in
run(config)
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\main.py", line 25, in run
model, ckpt_path = load_model(config, verbose=IS_RANK_ZERO)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\train\train_utils.py", line 228, in load_model
model = LightningTrainWrapper(config, verbose=verbose)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\train\trainer.py", line 37, in init
self.support_data = self.load_support_data()
^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\train\trainer.py", line 52, in load_support_data
support_data = generate_support_data(self.config, data_path=data_path, verbose=self.verbose)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\dataset\dataloader_factory.py", line 218, in generate_support_data
dset = TaskonomySegmentationDataset(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\dataset\taskonomy.py", line 547, in init
class_idxs = (torch.tensor(class_idxs)[idxs_perm[self.semseg_class]]).numpy()
~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
IndexError: index 117080 is out of bounds for dimension 0 with size 115595
The text was updated successfully, but these errors were encountered:
D:\OneDrive - Deakin University\Desktop\visual_token_matching>python main.py --stage 0 --task_fold 1
C:\Users\rajar\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\amp\autocast_mode.py:204: UserWarning: User provided device_type of 'cuda', but CUDA is not available. Disabling
warnings.warn('User provided device_type of 'cuda', but CUDA is not available. Disabling')
Load ckpt from model/pretrained_checkpoints\beit_base_patch16_224_pt22k.pth
Load state_dict by model_key = model
Expand the shared relative position embedding to each transformer block.
Traceback (most recent call last):
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\main.py", line 75, in
run(config)
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\main.py", line 25, in run
model, ckpt_path = load_model(config, verbose=IS_RANK_ZERO)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\train\train_utils.py", line 228, in load_model
model = LightningTrainWrapper(config, verbose=verbose)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\train\trainer.py", line 37, in init
self.support_data = self.load_support_data()
^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\train\trainer.py", line 52, in load_support_data
support_data = generate_support_data(self.config, data_path=data_path, verbose=self.verbose)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\dataset\dataloader_factory.py", line 218, in generate_support_data
dset = TaskonomySegmentationDataset(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\OneDrive - Deakin University\Desktop\visual_token_matching\dataset\taskonomy.py", line 547, in init
class_idxs = (torch.tensor(class_idxs)[idxs_perm[self.semseg_class]]).numpy()
~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
IndexError: index 117080 is out of bounds for dimension 0 with size 115595
The text was updated successfully, but these errors were encountered: