Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Clarity regarding training data #5

Closed
pra-dan opened this issue Aug 29, 2021 · 2 comments
Closed

Clarity regarding training data #5

pra-dan opened this issue Aug 29, 2021 · 2 comments

Comments

@pra-dan
Copy link

pra-dan commented Aug 29, 2021

I find much clarity lacking in the training process. Additional info on training dataset format is missing in toolkits/label_conversion/README.md. I understand that it will be update sometime soon.

The docs specify the training data to be formatted as:

# The id represent the correspondence relation
├─dataset root
│ ├─images/ id.jpg
│ ├─det_annotations/ id.json
│ ├─da_seg_annotations/ id.png
│ ├─ll_seg_annotations/ id.png

But the dataset downloaded from the bdd100k site has the following structure.

.
└── segmentation
    ├── __MACOSX
    │   └── test
    ├── test
    │   ├── __MACOSX
    │   │   └── test
    │   └── test
    │       └── raw_images
    ├── train
    │   ├── __MACOSX
    │   │   └── train
    │   └── train
    │       ├── class_color
    │       ├── class_id
    │       ├── instance_color
    │       ├── instance_id
    │       └── raw_images
    └── val
        ├── __MACOSX
        │   └── val
        └── val
            ├── class_color
            ├── class_id
            ├── instance_color
            ├── instance_id
            └── raw_images
  1. Its unclear which among instance_color, class_id and instance_id denote det_annotations,da_seg_annotations,
    ll_seg_annotations. All of them are masks. I dont' intend to use the object detection part, so the json conversion shouldn't be very necessary for now.

  2. The lib/config/default.py contains params such as

_C.DATASET.DATAROOT = '/home/zwt/bdd/bdd100k/images/100k'       # the path of images folder
_C.DATASET.LABELROOT = '/home/zwt/bdd/bdd100k/labels/100k'      # the path of det_annotations folder
_C.DATASET.MASKROOT = '/home/zwt/bdd/bdd_seg_gt'                # the path of da_seg_annotations folder
_C.DATASET.LANEROOT = '/home/zwt/bdd/bdd_lane_gt'

It would be better if more info can be provided for the paths such that it can be generalised.

@Riser6
Copy link
Collaborator

Riser6 commented Aug 30, 2021

I find much clarity lacking in the training process. Additional info on training dataset format is missing in toolkits/label_conversion/README.md. I understand that it will be update sometime soon.

The docs specify the training data to be formatted as:

# The id represent the correspondence relation
├─dataset root
│ ├─images/ id.jpg
│ ├─det_annotations/ id.json
│ ├─da_seg_annotations/ id.png
│ ├─ll_seg_annotations/ id.png

But the dataset downloaded from the bdd100k site has the following structure.

.
└── segmentation
    ├── __MACOSX
    │   └── test
    ├── test
    │   ├── __MACOSX
    │   │   └── test
    │   └── test
    │       └── raw_images
    ├── train
    │   ├── __MACOSX
    │   │   └── train
    │   └── train
    │       ├── class_color
    │       ├── class_id
    │       ├── instance_color
    │       ├── instance_id
    │       └── raw_images
    └── val
        ├── __MACOSX
        │   └── val
        └── val
            ├── class_color
            ├── class_id
            ├── instance_color
            ├── instance_id
            └── raw_images
  1. Its unclear which among instance_color, class_id and instance_id denote det_annotations,da_seg_annotations,
    ll_seg_annotations. All of them are masks. I dont' intend to use the object detection part, so the json conversion shouldn't be very necessary for now.
  2. The lib/config/default.py contains params such as
_C.DATASET.DATAROOT = '/home/zwt/bdd/bdd100k/images/100k'       # the path of images folder
_C.DATASET.LABELROOT = '/home/zwt/bdd/bdd100k/labels/100k'      # the path of det_annotations folder
_C.DATASET.MASKROOT = '/home/zwt/bdd/bdd_seg_gt'                # the path of da_seg_annotations folder
_C.DATASET.LANEROOT = '/home/zwt/bdd/bdd_lane_gt'

It would be better if more info can be provided for the paths such that it can be generalised.

Thanks for your suggestion for our project, we will make improvement for it as soon as possible! Hope you can continue to pay attention to our project!

@Riser6
Copy link
Collaborator

Riser6 commented Aug 30, 2021

I find much clarity lacking in the training process. Additional info on training dataset format is missing in toolkits/label_conversion/README.md. I understand that it will be update sometime soon.

The docs specify the training data to be formatted as:

# The id represent the correspondence relation
├─dataset root
│ ├─images/ id.jpg
│ ├─det_annotations/ id.json
│ ├─da_seg_annotations/ id.png
│ ├─ll_seg_annotations/ id.png

But the dataset downloaded from the bdd100k site has the following structure.

.
└── segmentation
    ├── __MACOSX
    │   └── test
    ├── test
    │   ├── __MACOSX
    │   │   └── test
    │   └── test
    │       └── raw_images
    ├── train
    │   ├── __MACOSX
    │   │   └── train
    │   └── train
    │       ├── class_color
    │       ├── class_id
    │       ├── instance_color
    │       ├── instance_id
    │       └── raw_images
    └── val
        ├── __MACOSX
        │   └── val
        └── val
            ├── class_color
            ├── class_id
            ├── instance_color
            ├── instance_id
            └── raw_images
  1. Its unclear which among instance_color, class_id and instance_id denote det_annotations,da_seg_annotations,
    ll_seg_annotations. All of them are masks. I dont' intend to use the object detection part, so the json conversion shouldn't be very necessary for now.
  2. The lib/config/default.py contains params such as
_C.DATASET.DATAROOT = '/home/zwt/bdd/bdd100k/images/100k'       # the path of images folder
_C.DATASET.LABELROOT = '/home/zwt/bdd/bdd100k/labels/100k'      # the path of det_annotations folder
_C.DATASET.MASKROOT = '/home/zwt/bdd/bdd_seg_gt'                # the path of da_seg_annotations folder
_C.DATASET.LANEROOT = '/home/zwt/bdd/bdd_lane_gt'

It would be better if more info can be provided for the paths such that it can be generalised.

I have uploaded our training label to Google Drive, which can be downloaded through the link in Readme.md. md. Thank you for your attention and suggestions to our project

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants