Papers
Topics
Authors
Recent
Search
2000 character limit reached

A slice classification neural network for automated classification of axial PET/CT slices from a multi-centric lymphoma dataset

Published 11 Mar 2024 in eess.IV, cs.CV, cs.LG, and physics.med-ph | (2403.07105v1)

Abstract: Automated slice classification is clinically relevant since it can be incorporated into medical image segmentation workflows as a preprocessing step that would flag slices with a higher probability of containing tumors, thereby directing physicians attention to the important slices. In this work, we train a ResNet-18 network to classify axial slices of lymphoma PET/CT images (collected from two institutions) depending on whether the slice intercepted a tumor (positive slice) in the 3D image or if the slice did not (negative slice). Various instances of the network were trained on 2D axial datasets created in different ways: (i) slice-level split and (ii) patient-level split; inputs of different types were used: (i) only PET slices and (ii) concatenated PET and CT slices; and different training strategies were employed: (i) center-aware (CAW) and (ii) center-agnostic (CAG). Model performances were compared using the area under the receiver operating characteristic curve (AUROC) and the area under the precision-recall curve (AUPRC), and various binary classification metrics. We observe and describe a performance overestimation in the case of slice-level split as compared to the patient-level split training. The model trained using patient-level split data with the network input containing only PET slices in the CAG training regime was the best performing/generalizing model on a majority of metrics. Our models were additionally more closely compared using the sensitivity metric on the positive slices from their respective test sets.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)
  1. Slattery, A., “Validating an image segmentation program devised for staging lymphoma,” Australas. Phys. Eng. Sci. Med. 40, 799–809 (Dec. 2017).
  2. Weisman, A. J., Kim, J., Lee, I., McCarten, K. M., Kessel, S., Schwartz, C. L., Kelly, K. M., Jeraj, R., Cho, S. Y., and Bradshaw, T. J., “Automated quantification of baseline imaging PET metrics on FDG PET/CT images of pediatric hodgkin lymphoma patients,” EJNMMI Phys. 7, 76 (Dec. 2020).
  3. Ahamed, S., Dubljevic, N., Bloise, I., Gowdy, C., Martineau, P., Wilson, D., Uribe, C. F., Rahmim, A., and Yousefirizi, F., “A cascaded deep network for automated tumor detection and segmentation in clinical PET imaging of diffuse large B-cell lymphoma,” in [Medical Imaging 2022: Image Processing ], Colliot, O. and Išgum, I., eds., 12032, 120323M, International Society for Optics and Photonics, SPIE (2022).
  4. Capobianco, N., Meignan, M., Cottereau, A.-S., Vercellino, L., Sibille, L., Spottiswoode, B., Zuehlsdorff, S., Casasnovas, O., Thieblemont, C., and Buvat, I., “Deep-learning 18F-FDG uptake classification enables total metabolic tumor volume estimation in diffuse large b-cell lymphoma,” J. Nucl. Med. 62, 30–36 (Jan. 2021).
  5. Cottereau, A.-S., Versari, A., Loft, A., Casasnovas, O., Bellei, M., Ricci, R., Bardet, S., Castagnoli, A., Brice, P., Raemaekers, J., Deau, B., Fortpied, C., Raveloarivahy, T., Van Zele, E., Chartier, L., Vander Borght, T., Federico, M., Hutchings, M., Ricardi, U., Andre, M., and Meignan, M., “Prognostic value of baseline metabolic tumor volume in early-stage hodgkin lymphoma in the standard arm of the H10 trial,” Blood 131, 1456–1463 (Mar. 2018).
  6. Blanc-Durand, P., Van Der Gucht, A., Schaefer, N., Itti, E., and Prior, J. O., “Automatic lesion detection and segmentation of 18F-FET PET in gliomas: A full 3D U-Net convolutional neural network study,” PLoS One 13, e0195798 (Apr. 2018).
  7. Ceriani, L., Milan, L., Martelli, M., Ferreri, A. J. M., Cascione, L., Zinzani, P. L., Di Rocco, A., Conconi, A., Stathis, A., Cavalli, F., Bellei, M., Cozens, K., Porro, E., Giovanella, L., Johnson, P. W., and Zucca, E., “Metabolic heterogeneity on baseline 18FDG-PET/CT scan is a predictor of outcome in primary mediastinal b-cell lymphoma,” Blood 132, 179–186 (July 2018).
  8. Ahamed, S., Toosi, A., Uribe, C., Rahmim, A., and Yousefirizi, F., “Towards enhanced automated tumor detection using background slice annotation methods in clinical pet imaging of lymphoma,” Journal of Nuclear Medicine 63(supplement 2), 3346–3346 (2022).
  9. Ahamed, S., Chaussé, G., Klyuzhin, I., Rahmim, A., and Yousefirizi, F., “A comparative study of tumor detection models trained on coronal versus sagittal versus axial pet imaging slices,” Journal of Nuclear Medicine 63(supplement 2), 3245–3245 (2022).
  10. Zhao, X., Li, L., Lu, W., and Tan, S., “Tumor co-segmentation in PET/CT using multi-modality fully convolutional neural network,” Phys. Med. Biol. 64, 015011 (Dec. 2018).
  11. Shen, D., Wu, G., and Suk, H.-I., “Deep learning in medical image analysis,” Annual Review of Biomedical Engineering 19(1), 221–248 (2017). PMID: 28301734.
  12. Haenssle, H. A., Fink, C., Schneiderbauer, R., Toberer, F., Buhl, T., Blum, A., Kalloo, A., Hassen, A. B. H., Thomas, L., Enk, A., Uhlmann, L., Reader study level-I and level-II Groups, Alt, C., Arenbergerova, M., Bakos, R., Baltzer, A., Bertlich, I., Blum, A., Bokor-Billmann, T., Bowling, J., Braghiroli, N., Braun, R., Buder-Bakhaya, K., Buhl, T., Cabo, H., Cabrijan, L., Cevic, N., Classen, A., Deltgen, D., Fink, C., Georgieva, I., Hakim-Meibodi, L.-E., Hanner, S., Hartmann, F., Hartmann, J., Haus, G., Hoxha, E., Karls, R., Koga, H., Kreusch, J., Lallas, A., Majenka, P., Marghoob, A., Massone, C., Mekokishvili, L., Mestel, D., Meyer, V., Neuberger, A., Nielsen, K., Oliviero, M., Pampena, R., Paoli, J., Pawlik, E., Rao, B., Rendon, A., Russo, T., Sadek, A., Samhaber, K., Schneiderbauer, R., Schweizer, A., Toberer, F., Trennheuser, L., Vlahova, L., Wald, A., Winkler, J., Wölbing, P., and Zalaudek, I., “Man against machine: diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists,” Ann. Oncol. 29, 1836–1842 (Aug. 2018).
  13. Ardila, D., Kiraly, A. P., Bharadwaj, S., Choi, B., Reicher, J. J., Peng, L., Tse, D., Etemadi, M., Ye, W., Corrado, G., Naidich, D. P., and Shetty, S., “End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography,” Nat. Med. 25, 954–961 (June 2019).
  14. Wu, N., Phang, J., Park, J., Shen, Y., Huang, Z., Zorin, M., Jastrzebski, S., Fevry, T., Katsnelson, J., Kim, E., Wolfson, S., Parikh, U., Gaddam, S., Lin, L. L. Y., Ho, K., Weinstein, J. D., Reig, B., Gao, Y., Toth, H., Pysarenko, K., Lewin, A., Lee, J., Airola, K., Mema, E., Chung, S., Hwang, E., Samreen, N., Kim, S. G., Heacock, L., Moy, L., Cho, K., and Geras, K. J., “Deep neural networks improve radiologists’ performance in breast cancer screening,” IEEE Trans. Med. Imaging 39, 1184–1194 (Apr. 2020).
  15. Yousefirizi, F., Dubljevic, N., Ahamed, S., Bloise, I., Gowdy, C., O, J. H., Farag, Y., de Schaetzen, R., Martineau, P., Wilson, D., Uribe, C. F., and Rahmim, A., “Convolutional neural network with a hybrid loss function for fully automated segmentation of lymphoma lesions in FDG PET images,” in [Medical Imaging 2022: Image Processing ], Colliot, O. and Išgum, I., eds., 12032, 120320V, International Society for Optics and Photonics, SPIE (2022).
  16. Warfield, S. K., Zou, K. H., and Wells, W. M., “Simultaneous truth and performance level estimation (STAPLE): an algorithm for the validation of image segmentation,” IEEE Trans. Med. Imaging 23, 903–921 (July 2004).
  17. Lev, M. and Gonzalez, R., “17 - ct angiography and ct perfusion imaging,” in [Brain Mapping: The Methods (Second Edition) ], Toga, A. W. and Mazziotta, J. C., eds., 427–484, Academic Press, San Diego, second edition ed. (2002).
  18. He, K., Zhang, X., Ren, S., and Sun, J., “Deep residual learning for image recognition,” (2015).
  19. Lin, T.-Y., Goyal, P., Girshick, R., He, K., and Dollár, P., “Focal loss for dense object detection,” (2017).
Citations (1)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.