You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi 👋, thanks for your great job! And I have some questions about the Text-based MLC to confirm.
When using MIMIC-CXR for pre-training, is the label of multi-label classification extracted by CheXpert labeler?
In the code(class pretrain_dataset), I found that all the data of MIMIC-CXR was used in the pre-training of multi-label classification. Will this not cause information leakage in the subsequent downstream tasks?
Thank you in advance. I am looking forward to hearing from you!
The text was updated successfully, but these errors were encountered:
the text-based MLC will not cause the information leakage, since 1) we only use the training split of MIMIC-CXR for pretraining, and 2) the pretraining text-based MLC task enforces the alignment of images and paired textual formed labels, where the labels play the similar role with the report. The downstream diagnosis classification task, however, performs multi-label classification based on one-hot formed label.
Thanks for your reply!
I would also like to confirm that: Is text-based MLC trained with the whole network in Fig. 8? (The multi-label classification performance based on the updated feature $f_g^{kv}$ can also be effective enhanced.) But why is it called a pretraining task?
Hi 👋, thanks for your great job! And I have some questions about the Text-based MLC to confirm.
class pretrain_dataset
), I found that all the data of MIMIC-CXR was used in the pre-training of multi-label classification. Will this not cause information leakage in the subsequent downstream tasks?Thank you in advance. I am looking forward to hearing from you!
The text was updated successfully, but these errors were encountered: