Self-supervised pretraining has shown great performance in improving the accuracy of downstream tasks. Although pretraining on a large dataset improves performances, it becomes challenging to further optimize the model by solely enlarging the dataset. In contrast, additional adaptation of pretrained models to the target domain has shown promise in NLP. Inspired by the success of continual pretraining, we investigated the efficacy of adapting the target domain dataset to a pretrained model in medical imaging, particularly in the context of segmentation. We present a study based on a self-supervised pretraining framework using the SwinUNETR backbone. In this study, we improved the generalizability of the self-supervised pretraining by adapting a foundational model pretrained on 5k CT volumes to data of the downstream segmentation task. In detail, we employed 385 abdominal CT volumes for the continual task-adaptive pretraining and 24 abdominal CT volumes for the downstream segmentation task, all sourced from the same dataset. Additionally, we conducted comparative experiments to demonstrate the benefits of this task-adapting pretraining approach. Our method has shown that continual pretraining helps to improve the performances, achieving an average Dice score for 10-class organ segmentation of 87.8%.
|