Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Dataset details #6

Open
Atik-Ahamed opened this issue Aug 27, 2024 · 5 comments
Open

Dataset details #6

Atik-Ahamed opened this issue Aug 27, 2024 · 5 comments

Comments

@Atik-Ahamed
Copy link

Dear Authors,
Excellent works!
In your paper, you mentioned using 2656 brain MRI images from four public datasets (ADNI, ABIDE, ADHD, IXI). However, those data sources have many more images than what you have used. I was wondering, how did you select particularly 2656 images. Can you please release those sets of images? Or, can you please provide guidance on how to download those particular sets of images?

Thanks in advance!

Repository owner deleted a comment from amir1387aht Aug 27, 2024
@MungoMeng
Copy link
Owner

Hi, thanks for your interests in our work!
We collected the training set in 2019, so the included datasets (ADNI, ABIDE, ADHD, IXI) were extended much in the past five years (and we did not update). There are also some scans being excluded as their skull stripping processing is unsuccessfully.
Actually, it's unnecessary to strictly follow our data settings. You can collect your own training set, including all your available images. Our method is not optimized for any specific datasets and could be directly used in your data.

@Atik-Ahamed
Copy link
Author

Thanks for mentioning more details.

@2250432015
Copy link

I am writing to seek clarification on the "train_pairs.npy" and "valid_pairs.npy" files associated with your work. Could you please provide insights into how the data pairs (train_pairs and valid_pairs) in these files were generated?If available, could you direct me to any detailed documentation or usage guidelines for these files
微信截图_20240908171107
微信截图_20240908171056

@MungoMeng
Copy link
Owner

Hi, thanks for your interest in our work!
Personally, I suggest you just adopt our code in network.py, which is the core technical contribution. The other code can be customed for different datasets and data structures.

In our settings, the "train_pairs.npy" and "valid_pairs.npy" are two files indicating which image pairs will be used for training and validation. Within the file is a list, such as [[filename_1, filename_2], [filename_3, filename_4], [filename_5, filename_6]], which includes three image pairs for training/validation. All these filenames_1/2/3/4/5/6 should be placed in the data_dir. In addition, the filename_1/2/3/4/5/6 have also been preprocessed as npz files, which can be directly loaded by np.load (Please refer to the code in datagenerator.py).

Overall, these data loading code is highly customed. That's the reason why I suggest you just use our network code in your own data-loading code framework.

@xxxh111
Copy link

xxxh111 commented Sep 9, 2024

hi,may I ask if you could provide detailed information on preprocessing, a more accurate process or code? Thank you very much.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants