You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for your leadership on the detectron/detectron2 development!
We recently noticed the new_baseline, and followed your instruction regarding your reply on "Improve documentation concerning the new config files #3225". However, when we train our data set with the following command line (note that we have 2 GPUs on a single machine. We use your code lazyconfig_train_net.py):
We got the following error message (I just wonder if you have met with the same error before, and what is the potential solution. ) :
[03/03 12:58:56 d2.utils.events]: eta: 17 days, 17:55:13 iter: 99 total_loss: 1.757 loss_cls: 0.05864 loss_box_reg: 0.0001416 loss_mask: 0.6881 loss_rpn_cls: 0.4922 loss_rpn_loc: 0.5106 time: 2.0937 data_time: 1.1836 lr: 0.0028296 max_mem: 19232M
Traceback (most recent call last):
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 189, in
launch(
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/launch.py", line 67, in launch
mp.spawn(
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 230, in spawn
return start_processes(fn, args, nprocs, join, daemon, start_method='spawn')
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 188, in start_processes
while not context.join():
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 150, in join
raise ProcessRaisedException(msg, error_index, failed_process.pid)
torch.multiprocessing.spawn.ProcessRaisedException:
-- Process 0 terminated with the following error:
Traceback (most recent call last):
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/data/catalog.py", line 51, in get
f = self[name]
File "/home/anaconda3/envs/pytorch2/lib/python3.9/collections/init.py", line 1058, in getitem
raise KeyError(key)
KeyError: ['targetPartSeg_test']
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 59, in _wrap
fn(i, *args)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/launch.py", line 126, in _distributed_worker
main_func(*args)
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 184, in main
do_train(args, cfg)
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 120, in do_train
trainer.train(start_iter, cfg.train.max_iter)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/train_loop.py", line 150, in train
self.after_step()
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/train_loop.py", line 180, in after_step
h.after_step()
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/hooks.py", line 552, in after_step
self._do_eval()
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/hooks.py", line 525, in _do_eval
results = self._func()
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 103, in
hooks.EvalHook(cfg.train.eval_period, lambda: do_test(cfg, model)),
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 44, in do_test
model, instantiate(cfg.dataloader.test), instantiate(cfg.dataloader.evaluator)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/config/instantiate.py", line 77, in instantiate
return cls(**cfg)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/evaluation/coco_evaluation.py", line 130, in init
convert_to_coco_json(dataset_name, cache_path)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/data/datasets/coco.py", line 470, in convert_to_coco_json
coco_dict = convert_to_coco_dict(dataset_name)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/data/datasets/coco.py", line 326, in convert_to_coco_dict
dataset_dicts = DatasetCatalog.get(dataset_name)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/data/catalog.py", line 53, in get
raise KeyError(
KeyError: "Dataset '['targetPartSeg_test']' is not registered! Available datasets are: coco_2014_train, coco_2014_val, coco_2014_minival, coco_2014_minival_100, coco_2014_valminusminival, coco_2017_train, coco_2017_val, coco_2017_test, coco_2017_test-dev, coco_2017_val_100, keypoints_coco_2014_train, keypoints_coco_2014_val, keypoints_coco_2014_minival, keypoints_coco_2014_valminusminival, keypoints_coco_2014_minival_100, keypoints_coco_2017_train, keypoints_coco_2017_val, keypoints_coco_2017_val_100, coco_2017_train_panoptic_separated, coco_2017_train_panoptic_stuffonly, coco_2017_train_panoptic, coco_2017_val_panoptic_separated, coco_2017_val_panoptic_stuffonly, coco_2017_val_panoptic, coco_2017_val_100_panoptic_separated, coco_2017_val_100_panoptic_stuffonly, coco_2017_val_100_panoptic, lvis_v1_train, lvis_v1_val, lvis_v1_test_dev, lvis_v1_test_challenge, lvis_v0.5_train, lvis_v0.5_val, lvis_v0.5_val_rand_100, lvis_v0.5_test, lvis_v0.5_train_cocofied, lvis_v0.5_val_cocofied, cityscapes_fine_instance_seg_train, cityscapes_fine_sem_seg_train, cityscapes_fine_instance_seg_val, cityscapes_fine_sem_seg_val, cityscapes_fine_instance_seg_test, cityscapes_fine_sem_seg_test, cityscapes_fine_panoptic_train, cityscapes_fine_panoptic_val, voc_2007_trainval, voc_2007_train, voc_2007_val, voc_2007_test, voc_2012_trainval, voc_2012_train, voc_2012_val, ade20k_sem_seg_train, ade20k_sem_seg_val, targetPartSeg_train, targetPartSeg_test"
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi @ppwwyyxx ,
Thank you for your leadership on the detectron/detectron2 development!
We recently noticed the new_baseline, and followed your instruction regarding your reply on "Improve documentation concerning the new config files #3225". However, when we train our data set with the following command line (note that we have 2 GPUs on a single machine. We use your code lazyconfig_train_net.py):
python3 lazyconfig_train_net.py --config-file /home/anaconda3/envs/detectron2_0.6_cp9/lib/python3.9/site-packages/detectron2/model_zoo/configs/new_baselines/mask_rcnn_R_50_FPN_100ep_LSJ.py --num-gpus 2
We got the following error message (I just wonder if you have met with the same error before, and what is the potential solution. ) :
[03/03 12:58:56 d2.utils.events]: eta: 17 days, 17:55:13 iter: 99 total_loss: 1.757 loss_cls: 0.05864 loss_box_reg: 0.0001416 loss_mask: 0.6881 loss_rpn_cls: 0.4922 loss_rpn_loc: 0.5106 time: 2.0937 data_time: 1.1836 lr: 0.0028296 max_mem: 19232M
Traceback (most recent call last):
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 189, in
launch(
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/launch.py", line 67, in launch
mp.spawn(
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 230, in spawn
return start_processes(fn, args, nprocs, join, daemon, start_method='spawn')
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 188, in start_processes
while not context.join():
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 150, in join
raise ProcessRaisedException(msg, error_index, failed_process.pid)
torch.multiprocessing.spawn.ProcessRaisedException:
-- Process 0 terminated with the following error:
Traceback (most recent call last):
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/data/catalog.py", line 51, in get
f = self[name]
File "/home/anaconda3/envs/pytorch2/lib/python3.9/collections/init.py", line 1058, in getitem
raise KeyError(key)
KeyError: ['targetPartSeg_test']
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 59, in _wrap
fn(i, *args)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/launch.py", line 126, in _distributed_worker
main_func(*args)
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 184, in main
do_train(args, cfg)
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 120, in do_train
trainer.train(start_iter, cfg.train.max_iter)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/train_loop.py", line 150, in train
self.after_step()
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/train_loop.py", line 180, in after_step
h.after_step()
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/hooks.py", line 552, in after_step
self._do_eval()
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/engine/hooks.py", line 525, in _do_eval
results = self._func()
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 103, in
hooks.EvalHook(cfg.train.eval_period, lambda: do_test(cfg, model)),
File "/home/Documents/targetFeature/new_config/lazyconfig_train_net.py", line 44, in do_test
model, instantiate(cfg.dataloader.test), instantiate(cfg.dataloader.evaluator)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/config/instantiate.py", line 77, in instantiate
return cls(**cfg)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/evaluation/coco_evaluation.py", line 130, in init
convert_to_coco_json(dataset_name, cache_path)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/data/datasets/coco.py", line 470, in convert_to_coco_json
coco_dict = convert_to_coco_dict(dataset_name)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/data/datasets/coco.py", line 326, in convert_to_coco_dict
dataset_dicts = DatasetCatalog.get(dataset_name)
File "/home/anaconda3/envs/pytorch2/lib/python3.9/site-packages/detectron2/data/catalog.py", line 53, in get
raise KeyError(
KeyError: "Dataset '['targetPartSeg_test']' is not registered! Available datasets are: coco_2014_train, coco_2014_val, coco_2014_minival, coco_2014_minival_100, coco_2014_valminusminival, coco_2017_train, coco_2017_val, coco_2017_test, coco_2017_test-dev, coco_2017_val_100, keypoints_coco_2014_train, keypoints_coco_2014_val, keypoints_coco_2014_minival, keypoints_coco_2014_valminusminival, keypoints_coco_2014_minival_100, keypoints_coco_2017_train, keypoints_coco_2017_val, keypoints_coco_2017_val_100, coco_2017_train_panoptic_separated, coco_2017_train_panoptic_stuffonly, coco_2017_train_panoptic, coco_2017_val_panoptic_separated, coco_2017_val_panoptic_stuffonly, coco_2017_val_panoptic, coco_2017_val_100_panoptic_separated, coco_2017_val_100_panoptic_stuffonly, coco_2017_val_100_panoptic, lvis_v1_train, lvis_v1_val, lvis_v1_test_dev, lvis_v1_test_challenge, lvis_v0.5_train, lvis_v0.5_val, lvis_v0.5_val_rand_100, lvis_v0.5_test, lvis_v0.5_train_cocofied, lvis_v0.5_val_cocofied, cityscapes_fine_instance_seg_train, cityscapes_fine_sem_seg_train, cityscapes_fine_instance_seg_val, cityscapes_fine_sem_seg_val, cityscapes_fine_instance_seg_test, cityscapes_fine_sem_seg_test, cityscapes_fine_panoptic_train, cityscapes_fine_panoptic_val, voc_2007_trainval, voc_2007_train, voc_2007_val, voc_2007_test, voc_2012_trainval, voc_2012_train, voc_2012_val, ade20k_sem_seg_train, ade20k_sem_seg_val, targetPartSeg_train, targetPartSeg_test"
Beta Was this translation helpful? Give feedback.
All reactions