Skip to main content
Top

2023 | Book

Head and Neck Tumor Segmentation and Outcome Prediction

Third Challenge, HECKTOR 2022, Held in Conjunction with MICCAI 2022, Singapore, September 22, 2022, Proceedings

Editors: Vincent Andrearczyk, Valentin Oreiller, Mathieu Hatt, Adrien Depeursinge

Publisher: Springer Nature Switzerland

Book Series : Lecture Notes in Computer Science

insite
SEARCH

About this book

This book constitutes the Third 3D Head and Neck Tumor Segmentation in PET/CT Challenge, HECKTOR 2022, which was held in conjunction with the 25th International Conference on Medical Image Computing and Computer-Assisted Intervention, MICCAI 2022, on September 22, 2022.

The 22 contributions presented, as well as an overview paper, were carefully reviewed and selected from 24 submissions. This challenge aims to evaluate and compare the current state-of-the-art methods for automatic head and neck tumor segmentation. In the context of this challenge, a dataset of 883 delineated PET/CT images was made available for training.

Table of Contents

Frontmatter
Overview of the HECKTOR Challenge at MICCAI 2022: Automatic Head and Neck Tumor Segmentation and Outcome Prediction in PET/CT
Abstract
This paper presents an overview of the third edition of the HEad and neCK TumOR segmentation and outcome prediction (HECKTOR) challenge, organized as a satellite event of the 25th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI) 2022. The challenge comprises two tasks related to the automatic analysis of FDG-PET/CT images for patients with Head and Neck cancer (H &N), focusing on the oropharynx region. Task 1 is the fully automatic segmentation of H &N primary Gross Tumor Volume (GTVp) and metastatic lymph nodes (GTVn) from FDG-PET/CT images. Task 2 is the fully automatic prediction of Recurrence-Free Survival (RFS) from the same FDG-PET/CT and clinical data. The data were collected from nine centers for a total of 883 cases consisting of FDG-PET/CT images and clinical information, split into 524 training and 359 test cases. The best methods obtained an aggregated Dice Similarity Coefficient (\(DSC_{agg}\)) of 0.788 in Task 1, and a Concordance index (C-index) of 0.682 in Task 2.
Vincent Andrearczyk, Valentin Oreiller, Moamen Abobakr, Azadeh Akhavanallaf, Panagiotis Balermpas, Sarah Boughdad, Leo Capriotti, Joel Castelli, Catherine Cheze Le Rest, Pierre Decazes, Ricardo Correia, Dina El-Habashy, Hesham Elhalawani, Clifton D. Fuller, Mario Jreige, Yomna Khamis, Agustina La Greca, Abdallah Mohamed, Mohamed Naser, John O. Prior, Su Ruan, Stephanie Tanadini-Lang, Olena Tankyevych, Yazdan Salimi, Martin Vallières, Pierre Vera, Dimitris Visvikis, Kareem Wahid, Habib Zaidi, Mathieu Hatt, Adrien Depeursinge
Automated Head and Neck Tumor Segmentation from 3D PET/CT HECKTOR 2022 Challenge Report
Abstract
Head and neck tumor segmentation challenge (HECKTOR) 2022 offers a platform for researchers to compare their solutions to segmentation of tumors and lymph nodes from 3D CT and PET images. In this work, we describe our solution to HECKTOR 2022 segmentation task. We re-sample all images to a common resolution, crop around head and neck region, and train SegResNet semantic segmentation network from MONAI. We use 5-fold cross validation to select best model checkpoint. The final submission is an ensemble of 15 models from 3 runs. Our solution (team name NVAUTO) achieves the 1st place on the HECKTOR22 challenge leaderboard with an aggregated dice score of 0.78802 (https://​hecktor.​grand-challenge.​org/​evaluation/​segmentation/​leaderboard/​.). It is implemented with Auto3DSeg (https://​monai.​io/​apps/​auto3dseg.).
Andriy Myronenko, Md Mahfuzur Rahman Siddiquee, Dong Yang, Yufan He, Daguang Xu
A Coarse-to-Fine Ensembling Framework for Head and Neck Tumor and Lymph Segmentation in CT and PET Images
Abstract
Head and neck (H &N) cancer is one of the most prevalent cancers [1]. In its treatment and prognosis analysis, tumors and metastatic lymph nodes may play an important role but their manual segmentations are time-consuming and laborious. In this paper, we propose a coarse-to-fine ensembling framework to segment the H &N tumor and metastatic lymph nodes automatically from Positron Emission Tomography (PET) and Computed Tomography (CT) images. The framework consists of three steps. The first step is to locate the head region in CT images. The second step is a coarse segmentation, to locate the tumor and lymph region of interest (ROI) from the head region. The last step is a fine segmentation, to get the final precise predictions of tumors and metastatic lymph nodes, where we proposed a ensembling refinement model. This framework is evaluated quantitatively with aggregated Dice Similarity Coefficient (DSC) of 0.77782 in the task 1 of the HECKTOR 2022 challenge[2, 3] as team SJTU426.
Xiao Sun, Chengyang An, Lisheng Wang
A General Web-Based Platform for Automatic Delineation of Head and Neck Gross Tumor Volumes in PET/CT Images
Abstract
Delineation of head and neck lesions are crucial for radiation treatment planning and follow-up studies. In this paper we developed an automated segmentation method for head and neck primary and nodal gross tumor volumes (GTVp and GTVn) segmentation in positron emission tomography/computed tomography (PET/CT) provided by the MICCAI 2022 Head and Neck Tumor Segmentation Challenge (HECKTOR 2022). Our segmentation algorithm takes nnU-Net as the backbone and uses dedicated pre- and post-processing to improve the auto-segmentation performance. The pipeline described achieved DSC results of 0.77212 (GTVp 0.77485 and GTVn 0.76938) in the testing dataset of HECTOR 2022. The developed auto-segmentation method is further extensively developed to a web-based platform to permit easy access and facilitate clinical workflow.
Hao Jiang, Jason Haimerl, Xuejun Gu, Weiguo Lu
Octree Boundary Transfiner: Efficient Transformers for Tumor Segmentation Refinement
Abstract
In this paper, we create a fully autonomous system that segments primary head and neck tumors as well as lymph node tumors given only FDG-PET and CT scans without contrast enhancers. Given only these two modalities, the typical Dice score for the state-of-the-art (SOTA) models lies below 0.8, below what it would be when including other modalities due to the low resolution of PET scans and noisy non-enhanced CT images. Thus, we seek to improve tumor segmentation accuracy while working with the limitation of only having these two modalities. We introduce the Transfiner, a novel octree-based refinement system to harness the fidelity of transformers while keeping computation and memory costs low for fast inferencing. The observation behind our method is that segmentation errors almost always occur at the edges of a mask for predictions from a well-trained model. The Transfiner utilizes base network feature maps in addition to the raw modalities as input and selects regions of interest from these. These are then processed with a transformer network and decoded with a CNN. We evaluated our framework with Dice Similarity Coefficient (DSC) 0.76426 for the first task of the Head and Neck Tumor Segmentation Challenge (HECKTOR) and ranked 6th.
Anthony Wang, Ti Bai, Dan Nguyen, Steve Jiang
Head and Neck Primary Tumor and Lymph Node Auto-segmentation for PET/CT Scans
Abstract
Segmentation of head and neck (H &N) cancer primary tumor and lymph nodes on medical imaging is a routine part of radiation treatment planning for patients and may lead to improved response assessment and quantitative imaging analysis. Manual segmentation is a difficult and time-intensive task, requiring specialist knowledge. In the area of computer vision, deep learning-based architectures have achieved state-of-the-art (SOTA) performances for many downstream tasks, including medical image segmentation. Deep learning-based auto-segmentation tools may improve efficiency and robustness of H &N cancer segmentation. For the purpose of encouraging high performing methods for lesion segmentation while utilizing the bi-modal information of PET and CT images, the HEad and neCK TumOR (HECKTOR) challenge is offered annually. In this paper, we preprocess PET/CT images and train and evaluate several deep learning frameworks, including 3D U-Net, MNet, Swin Transformer, and nnU-Net (both 2D and 3D), to segment CT and PET images of primary tumors (GTVp) and cancerous lymph nodes (GTVn) automatically. Our investigations led us to three promising models for submission. Via 5-fold cross validation with ensembling and testing on a blinded hold-out set, we received an average of 0.77 and 0.70 using the aggregated Dice Similarity Coefficient (DSC) metric for primary and node, respectively, for task 1 of the HECKTOR2022 challenge. Herein, we describe in detail the methodology and results for our top three performing models that were submitted to the challenge. Our investigations demonstrate the versatility and robustness of such deep learning models on automatic tumor segmentation to improve H &N cancer treatment. Our full implementation based on the PyTorch framework and the trained models are available at https://​github.​com/​xmuyzz/​HECKTOR2022 (Team name: AIMERS).
Arnav Jain, Julia Huang, Yashwanth Ravipati, Gregory Cain, Aidan Boyd, Zezhong Ye, Benjamin H. Kann
Fusion-Based Automated Segmentation in Head and Neck Cancer via Advance Deep Learning Techniques
Abstract
Background: Accurate prognostic stratification and segmentation of Head-and-Neck Squamous-Cell-Carcinoma (HNSCC) patients can be an important clinical reference when designing therapeutic strategies. We set to automatically segment HNSCC using advanced deep learning techniques linked to the image fusion technique.
Method: 883 subjects were extracted from HECKTOR-Challenge. 524 subjects were considered for the training and validation procedure, and 359 subjects as external testing were employed to validate our segmentation models. First, PET images were registered to CT images. The resultant images are cropped after an enhancement procedure. Subsequently, a weighted fusion technique was employed to combine PET and CT information. To this end, we developed a Cascade-Net consisting of two states of art neural networks to segment the tumors via the fused image. Our segmentation framework performs in three main stages. In the first stage, which is an organ localizer module, a candidate segmentation region of interest (ROIs) for each organ is generated. The second stage is a 3D U-Net refinement organ segmentation which produces a more robust and accurate contour from the previous coarse segmentation mask. This network is equipped with an attention mechanism on skip connections and a deep supervision concept that generates ROIs by eliminating irrelevant background information. This network will identify the probability of the presence of each organ. In the last stage, the extracted regions will be fed to the 3D ResU-Net to generate a fine segmentation. The performance of the proposed framework was evaluated through well-established quantitative metrics such as the dice similarity coefficient.
Result: Using the weighted fusion technique linked with Cascade-Net, our method provided the average dice score of 0.71. Moreover, this algorithm resulted in dice score of 0.74, and 0.68 for the primary gross tumor volume (GTVp) and metastatic nodes (GTVn), respectively.
Conclusion: We demonstrated that using the fusion technique followed by an appropriate automatic segmentation technique provides a good performance.
Seyed Masoud Rezaeijo, Ali Harimi, Mohammad R. Salmanpour
Stacking Feature Maps of Multi-scaled Medical Images in U-Net for 3D Head and Neck Tumor Segmentation
Abstract
Machine learning, especially deep learning, has achieved state-of-the-art performance on various computer vision tasks. For computer vision tasks in the medical domain, it remains as challenging tasks since medical data is heterogeneous, multi-level, and multi-scale. Head and Neck Tumor Segmentation Challenge (HECKTOR) provides a platform to apply machine learning techniques to the medical image domain. HECKTOR 2022 provides positron emission tomography/computed tomography (PET/CT) images which includes useful metabolic and anatomical information to sufficiently make an accurate tumor segmentation. In this paper, we proposed a stacked-multi-scaled medical image segmentation framework to automatically segment the Head and Neck tumor using PET/CT images. The main idea of our network was to generate various low-resolution feature maps of PET/CT images to make a better contour of Head and Neck tumors. We used multi-scaled PET/CT images as inputs, and stacked different intermediate feature maps by resolution for a better inference result. In addition, we evaluated our model on the HECKTOR challenge test dataset. Overall, we achieved a 0.69786, 0.66730 mean Dice score on GTVp and GTVn respectively. Our team’s name is HPCAS.
Yaying Shi, Xiaodong Zhang, Yonghong Yan
A Fine-Tuned 3D U-Net for Primary Tumor and Affected Lymph Nodes Segmentation in Fused Multimodal Images of Oropharyngeal Cancer
Abstract
Head and Neck (HN) cancer has the sixth highest incidence rate of all malignancies worldwide. One of the two main curative treatments for this malignancy is radiotherapy, whose delivery depends on accurate contouring of the primary tumor and affected lymph nodes among other structures. In this study, we present a transfer learning-based approach for the automatic primary tumor and lymph nodes segmentation in fused positron emission tomography (PET) and computed tomography (CT) images belonging to the HECKTOR challenge dataset. Transfer learning is performed from the Genesis Chest CT model, a publicly available 3D U-net, pre-trained on chest CT scans. Three-fold cross-validation is employed during training, so that, on each fold, two different binary segmentation models are chosen, one for the primary tumor and one for the lymph nodes. During testing, majority voting is applied. Our results show promising performance on the training and validation cohorts, while moderate performance was observed in the test cohort.
Agustina La Greca Saint-Esteven, Laura Motisi, Panagiotis Balermpas, Stephanie Tanadini-Lang
A U-Net Convolutional Neural Network with Multiclass Dice Loss for Automated Segmentation of Tumors and Lymph Nodes from Head and Neck Cancer PET/CT Images
Abstract
We implemented a 2D U-Net model with an ImageNet-pretrained ResNet50 encoder for performing segmentation of primary tumors (GTVp) and metastatic lymph nodes (GTVn) from PET/CT images provided by the HEad and neCK TumOR segmentation challenge (HECKTOR) 2022. We utilized a multiclass Dice Loss for model training which was minimized using the AMSGrad variant of the Adam algorithm optimizer. We trained our 2D models on the axial slices of the images in a 5-fold cross-validation setting and stacked the 2D predictions axially to obtain the predicted 3D segmentation masks. We obtained mean aggregate Dice similarity coefficients (mean DSC\(_{\text {agg}}\)) of 0.6865, 0.6689, 0.6768, 0.6792, and 0.6726 on the 5 validation sets respectively. The model with the best performance on the validation set (validation split 1) was chosen for evaluating segmentation masks on the test set for submission to the challenge. Our model achieved a mean DSC\(_{\text {agg}}\) = 0.6345 on the test set, with DSC\(_{\text {agg}}\)(GTVp) = 0.6955 and DSC\(_{\text {agg}}\)(GTVn) = 0.5734. The implementation can be found under our Github repository.
Shadab Ahamed, Luke Polson, Arman Rahmim
Multi-scale Fusion Methodologies for Head and Neck Tumor Segmentation
Abstract
Head and Neck (H &N) organ-at-risk (OAR) and tumor segmentations are an essential component of radiation therapy planning. The varying anatomic locations and dimensions of H &N nodal Gross Tumor Volumes (GTVn) and H &N primary gross tumor volume (GTVp) are difficult to obtain due to lack of accurate and reliable delineation methods. The downstream effect of incorrect segmentation can result in unnecessary irradiation of normal organs. Towards a fully automated radiation therapy planning algorithm, we explore the efficacy of multi-scale fusion based deep learning architectures for accurately segmenting H &N tumors from medical scans. Team Name: M &H_lab_NU.
Abhishek Srivastava, Debesh Jha, Bulent Aydogan, Mohamed E. Abazeed, Ulas Bagci
Swin UNETR for Tumor and Lymph Node Segmentation Using 3D PET/CT Imaging: A Transfer Learning Approach
Abstract
Delineation of Gross Tumor Volume (GTV) is essential for the treatment of cancer with radiotherapy. GTV contouring is a time-consuming specialized manual task performed by radiation oncologists. Deep Learning (DL) algorithms have shown potential in creating automatic segmentations, reducing delineation time and inter-observer variation. The aim of this work was to create automatic segmentations of primary tumors (GTVp) and pathological lymph nodes (GTVn) in oropharyngeal cancer patients using DL. The organizers of the HECKTOR 2022 challenge provided 3D Computed Tomography (CT) and Positron Emission Tomography (PET) scans with ground-truth GTV segmentations acquired from nine different centers. Bounding box cropping was applied to obtain an anatomic based region of interest. We used the Swin UNETR model in combination with transfer learning. The Swin UNETR encoder weights were initialized by pre-trained weights of a self-supervised Swin UNETR model. An average Dice score of 0.656 was achieved on a test set of 359 patients from the HECKTOR 2022 challenge. Code is available at: https://​github.​com/​HC94/​swin_​unetr_​hecktor_​2022.
Aicrowd Group Name: RT_UMCG
Hung Chu, Luis Ricardo De la O Arévalo, Wei Tang, Baoqiang Ma, Yan Li, Alessia De Biase, Stefan Both, Johannes Albertus Langendijk, Peter van Ooijen, Nanna Maria Sijtsema, Lisanne V. van Dijk
Simplicity Is All You Need: Out-of-the-Box nnUNet Followed by Binary-Weighted Radiomic Model for Segmentation and Outcome Prediction in Head and Neck PET/CT
Abstract
Automated lesion detection and segmentation might assist radiation therapy planning and contribute to the identification of prognostic image-based biomarkers towards personalized medicine. In this paper, we propose a pipeline to segment the primary and metastatic lymph nodes from fluorodeoxyglucose (FDG) positron emission tomography and computed tomography (PET/CT) head and neck (H &N) images and then predict recurrence free survival (RFS) based on the segmentation results. For segmentation, an out-of-the-box nnUNet-based deep learning method was trained and labelled the two lesion types as primary gross tumor volume (GTVp) and metastatic nodes (GTVn). For RFS prediction, 2421 radiomic features were extracted from the merged GTVp and GTVn using the pyradiomics package. The ability of each feature to predict RFS was measured using the C-index. Only the features with a C-index greater than \(C_{min}\), hyperparameter of the model, were selected and assigned a +1 or –1 weight as a function of how they varied with the recurrence time. The final RFS probability was calculated as the mean across all selected feature z-scores weighted by their +/–1 weight. The fully automated pipeline was applied to the data provided through the HECKTOR 2022 MICCAI challenge. On the test data, the fully automated segmentation model achieved 0.777 and 0.763 Dice scores on the primary tumor and lymph nodes respectively (0.770 on average). The binary-weighted radiomic model yielded a 0.682 C-index. These results allowed us to rank first for outcome prediction and fourth for segmentation in the challenge. We conclude that the proposed fully-automated pipeline from segmentation to outcome prediction using a binary-weighted radiomic model competes well with more complicated models. Team: LITO.
Louis Rebaud, Thibault Escobar, Fahad Khalid, Kibrom Girum, Irène Buvat
Radiomics-Enhanced Deep Multi-task Learning for Outcome Prediction in Head and Neck Cancer
Abstract
Outcome prediction is crucial for head and neck cancer patients as it can provide prognostic information for early treatment planning. Radiomics methods have been widely used for outcome prediction from medical images. However, these methods are limited by their reliance on intractable manual segmentation of tumor regions. Recently, deep learning methods have been proposed to perform end-to-end outcome prediction so as to remove the reliance on manual segmentation. Unfortunately, without segmentation masks, these methods will take the whole image as input, such that makes them difficult to focus on tumor regions and potentially unable to fully leverage the prognostic information within the tumor regions. In this study, we propose a radiomics-enhanced deep multi-task framework for outcome prediction from PET/CT images, in the context of HEad and neCK TumOR segmentation and outcome prediction challenge (HECKTOR 2022). In our framework, our novelty is to incorporate radiomics as an enhancement to our recently proposed Deep Multi-task Survival model (DeepMTS). The DeepMTS jointly learns to predict the survival risk scores of patients and the segmentation masks of tumor regions. Radiomics features are extracted from the predicted tumor regions and combined with the predicted survival risk scores for final outcome prediction, through which the prognostic information in tumor regions can be further leveraged. Our method achieved a C-index of 0.681 on the testing set, placing the 2nd on the leaderboard with only 0.00068 lower in C-index than the 1st place.
Mingyuan Meng, Lei Bi, Dagan Feng, Jinman Kim
Recurrence-Free Survival Prediction Under the Guidance of Automatic Gross Tumor Volume Segmentation for Head and Neck Cancers
Abstract
For Head and Neck Cancers (HNC) patient management, automatic gross tumor volume (GTV) segmentation and accurate pre-treatment cancer recurrence prediction are of great importance to assist physicians in designing personalized management plans, which have the potential to improve the treatment outcome and quality of life for HNC patients. In this paper, we developed an automated primary tumor (GTVp) and lymph nodes (GTVn) segmentation method based on combined pre-treatment positron emission tomography/computed tomography (PET/CT) scans of HNC patients. We extracted radiomics features from the segmented tumor volume and constructed a multi-modality tumor recurrence-free survival (RFS) prediction model, which fused the prediction results from separate CT radiomics, PET radiomics, and clinical models. We performed 5-fold cross-validation to train and evaluate our methods on the MICCAI 2022 HEad and neCK TumOR segmentation and outcome prediction challenge (HECKTOR) dataset. The ensemble prediction results on the testing cohort achieved aggregated Dice scores of 0.77 and 0.73 for GTVp and GTVn segmentation, respectively, and a C-index value of 0.67 for RFS prediction. The code is publicly available (https://​github.​com/​wangkaiwan/​HECKTOR-2022-AIRT). Our team’s name is AIRT.
Kai Wang, Yunxiang Li, Michael Dohopolski, Tao Peng, Weiguo Lu, You Zhang, Jing Wang
Joint nnU-Net and Radiomics Approaches for Segmentation and Prognosis of Head and Neck Cancers with PET/CT Images
Abstract
Automatic segmentation of head and neck cancer (HNC) tumors and lymph nodes plays a crucial role in the optimization treatment strategy and prognosis analysis. This study aims to employ nnU-Net for automatic segmentation and radiomics for recurrence-free survival (RFS) prediction using pretreatment PET/CT images in a multi-center HNC cohort of 883 patients (524 patients for training, 359 for testing) provided within the context of the HECKTOR MICCAI challenge 2022. A bounding box of the extended oropharyngeal region was retrieved for each patient with fixed size of 224 \(\times \) 224 \(\times \) 224 mm\(^{3}\). Then the 3D nnU-Net architecture was adopted to carry out automatic segmentation of both primary tumor and lymph nodes. From the predicted segmentation mask, ten conventional features and 346 standardized radiomics features were extracted for each patient. Three prognostic models were constructed containing conventional and radiomics features alone, and their combinations by multivariate CoxPH modelling. The statistical harmonization method, ComBat, was explored towards reducing multicenter variations. Dice score and C-index were used as evaluation metrics for segmentation and prognosis task, respectively. For segmentation task, we achieved a mean dice score of 0.7 for primary tumor and lymph nodes. For recurrence-free survival prediction, conventional and radiomics models obtained C-index values of 0.66 and 0.65 in the test set, respectively, while the combined model did not improve the prognostic performance (0.65).
Hui Xu, Yihao Li, Wei Zhao, Gwenolé Quellec, Lijun Lu, Mathieu Hatt
MLC at HECKTOR 2022: The Effect and Importance of Training Data When Analyzing Cases of Head and Neck Tumors Using Machine Learning
Abstract
Head and neck cancers are the fifth most common cancer worldwide, and recently, analysis of Positron Emission Tomography (PET) and Computed Tomography (CT) images has been proposed to identify patients with a prognosis. Even though the results look promising, more research is needed to further validate and improve the results. This paper presents the work done by team MLC for the 2022 version of the HECKTOR grand challenge held at MICCAI 2022. For Task 1, the automatic segmentation task, our approach was, in contrast to earlier solutions using 3D segmentation, to keep it as simple as possible using a 2D model, analyzing every slice as a standalone image. In addition, we were interested in understanding how different modalities influence the results. We proposed two approaches; one using only the CT scans to make predictions and another using a combination of the CT and PET scans. For Task 2, the prediction of recurrence-free survival, we first proposed two approaches, one where we only use patient data and one where we combined the patient data with segmentations from the image model. For the prediction of the first two approaches, we used Random Forest. In our third approach, we combined patient data and image data using XGBoost. Low kidney function might worsen cancer prognosis. In this approach, we therefore estimated the kidney function of the patients and included it as a feature. Overall, we conclude that our simple methods were not able to compete with the highest-ranking submissions, but we still obtained reasonably good scores. We also got interesting insights into how the combination of different modalities can influence the segmentation and predictions.
Vajira Thambawita, Andrea M. Storås, Steven A. Hicks, Pål Halvorsen, Michael A. Riegler
Towards Tumour Graph Learning for Survival Prediction in Head & Neck Cancer Patients
Abstract
With nearly one million new cases diagnosed worldwide in 2020, head & neck cancer is a deadly and common malignity. There are challenges to decision making and treatment of such cancer, due to lesions in multiple locations and outcome variability between patients. Therefore, automated segmentation and prognosis estimation approaches can help ensure each patient gets the most effective treatment. This paper presents a framework to perform these functions on arbitrary field of view (FoV) PET and CT registered scans, thus approaching tasks 1 and 2 of the HECKTOR 2022 challenge as team VokCow. The method consists of three stages: localization, segmentation and survival prediction. First, the scans with arbitrary FoV are cropped to the head and neck region and a u-shaped convolutional neural network (CNN) is trained to segment the region of interest. Then, using the obtained regions, another CNN is combined with a support vector machine classifier to obtain the semantic segmentation of the tumours, which results in an aggregated Dice score of 0.57 in task 1. Finally, survival prediction is approached with an ensemble of Weibull accelerated failure times model and deep learning methods. In addition to patient health record data, we explore whether processing graphs of image patches centred at the tumours via graph convolutions can improve the prognostic predictions. A concordance index of 0.64 was achieved in the test set, ranking 6th in the challenge leaderboard for this task.
Ángel Víctor Juanco-Müller, João F. C. Mota, Keith Goatman, Corné Hoogendoorn
Combining nnUNet and AutoML for Automatic Head and Neck Tumor Segmentation and Recurrence-Free Survival Prediction in PET/CT Images
Abstract
Head and neck (H &N) tumor segmentation from FDG-PET/CT images has a significant impact on radiotherapy diagnosis. However, manual delineation of primary tumor and lymph nodes is a time-consuming and labor-intensive process. Also, patients that underwent radiotherapy have a high risk of regional recurrence. In this work, we used 3D nnU-Net with DiceTopK loss function to achieve automatic segmentation for head and neck primary tumor and lymph nodes. With the average ensemble of five cross-validation models, our approach got an aggregated Dice Similarity Coefficient of 0.70427 on the test set. Furthermore, we extracted radiomics features from PET/CT images combined with the clinical data to predict patients’ Recurrence-Free Survival (RFS) using a weighted ensemble predictor by AutoGluon. This led to a Concordance index of 0.63896 on the test set. The code is publicly available at  https://​github.​com/​amylyu1123/​HECKTOR-2022.
Qing Lyu
Head and Neck Cancer Localization with Retina Unet for Automated Segmentation and Time-To-Event Prognosis from PET/CT Images
Abstract
Auto-segmentation of the primary tumor (GTVp) and the associated lymph nodes (GTVn) of head and neck cancer (HNC) is deemed beneficial for precise radiotherapy. However, previous studies were mostly focusing on the auto segmentation of GTVp. It remains difficult to additionally segment GTVn automatically. Moreover, current methods also face challenges in handling whole-body scans due to memory limitations. In this study, the Retina Unet has been utilized for the first time to localize the HNC from whole-body positron emission tomography/computed tomography (PET/CT) scans. Cropped (based on the predicted tumor center) PET/CTs were then input to a multi-label Unet to produce the auto segmentation of GTVp and GTVn. Several time-to-event analysis models, including a segmentation-free model, have also been explored for relapse free survival (RFS) prognosis. The proposed models achieved encouraging cross validation (testing) Dice coefficient scores for GTVp/GTVn segmentation of 0.66 (0.70), indicating a promising first step towards fully automated HNC localization and segmentation from whole-body PET/CT scans. Furthermore, the segmentation-free PET-only RFS prognosis model produced the best average cross-validation (testing) Harrell’s Concordance Index of 0.70 (0.635), verifying our previous observation that GTV segmentation might be less relevant for PET-based prognosis.
Yiling Wang, Elia Lombardo, Lili Huang, Claus Belka, Marco Riboldi, Christopher Kurz, Guillaume Landry
HNT-AI: An Automatic Segmentation Framework for Head and Neck Primary Tumors and Lymph Nodes in FDG- PET/CT Images
Abstract
Head and neck cancer is one of the most prevalent cancers in the world. Automatic delineation of primary tumors and lymph nodes is important for cancer diagnosis and treatment. In this paper, we develop a deep learning-based model for automatic tumor segmentation, HNT-AI, using PET/CT images provided by the MICCAI 2022 Head and Neck Tumor (HECKTOR) segmentation Challenge. We investigate the effect of residual blocks, squeeze-and-excitation normalization, and grid-attention gates on the performance of 3D-UNET. We project the predicted masks on the z-axis and apply k-means clustering to reduce the number of false positive predictions. Our proposed HNT-AI segmentation framework achieves an aggregated dice score of 0.774 and 0.759 for primary tumors and lymph nodes, respectively, on the unseen external test set. Qualitative analysis of the predicted segmentation masks shows that the predicted segmentation mask tends to follow the high standardized uptake value (SUV) area on the PET scans more closely than the ground truth masks. The largest tumor volume, the larget lymph node volume, and the total number of lymph nodes derived from the segmentation proved to be potential biomarkers for recurrence-free survival with a C-index of 0.627 on the test set.
Zohaib Salahuddin, Yi Chen, Xian Zhong, Nastaran Mohammadian Rad, Henry C. Woodruff, Philippe Lambin
Head and Neck Tumor Segmentation with 3D UNet and Survival Prediction with Multiple Instance Neural Network
Abstract
Head and Neck Squamous Cell Carcinoma (HNSCC) is a group of malignancies arising in the squamous cells of the head and neck region. As a group, HNSCC accounts for around 4.5% of cancer incidences and deaths worldwide. Radiotherapy is part of the standard care for HNSCC cancers and accurate delineation of tumors is important for treatment quality. Imaging features of Computed Tomography (CT) and Positron Emission Tomography (PET) scans have been shown to be correlated with survival of HNSCC patients. In this paper we present our solutions to the segmentation task and recurrence-free survival prediction task of the HECKTOR 2022 challenge. We trained a 3D UNet model for the segmentation of primary tumors and lymph node metastases based on CT images. Three sets of models with different combinations of loss functions were ensembled to generate a more robust model. The softmax output of the ensembled model was fused with co-registered PET scans and post-processed to generate our submission to task 1 of the challenge, which achieved a 0.716 aggregated Dice score on the test data. Our segmentation model outputs were used to extract radiomic features of individual tumors on test data. Clinical variables and location of the tumors were also encoded and concatenated with radiomic features as additional inputs. We trained a multiple instance neural network to aggregate features of individual tumors into patient-level representations and predict recurrence-free survival rates of patients. Our method achieved an AUC of 0.619 for task 2 on the test data (Team name: SMIAL).
Jianan Chen, Anne L. Martel
Deep Learning and Machine Learning Techniques for Automated PET/CT Segmentation and Survival Prediction in Head and Neck Cancer
Abstract
Background: Accurate prognostic stratification as well as segmentation of Head-and-Neck Squamous-Cell-Carcinoma (HNSCC) patients can be an important clinical reference when designing therapeutic strategies. We set to enable automated segmentation of tumors and prediction of recurrence-free survival (RFS) using advanced deep learning techniques and Hybrid Machine Learning Systems (HMLSs).
Method: In this work, 883 subjects were extracted from HECKTOR-Challenge: ~60% of the total subjects were considered for the training and validation procedure, and the remaining subjects for external testing were employed to validate our models. PET images were registered to CT. First, a weighted fusion technique was employed to combine PET and CT information. We also employed Cascade-Net to enable automated segmentation of HNSCC tumors. Moreover, we extracted deep learning features (DF) via a 3D auto-encoder algorithm from PET and the fused image. Subsequently, we employed an HMLS including a feature selection algorithm such as Mutual Information (MI) linked with a survival prediction algorithm such as Random Survival Forest (RSF) optimized by 5-fold cross-validation and grid search. The dataset with DFs was normalized by the z-score technique. Moreover, dice score and c-Index were reported to evaluate the segmentation and prediction models, respectively.
Result: For segmentation, the weighted fusion technique followed by the Cascade-Net segmentation algorithm resulted in a validation dice score of 72%. External testing of 71% confirmed our findings. DFs extracted from sole PET and MI followed by RSF enabled us to receive a validation c-index of 66% for RFS prediction. The external testing of 59% confirmed our finding.
Conclusion: We demonstrated that using the fusion technique followed by an appropriate automated segmentation technique provides good performance. Moreover, employing DFs extracted from sole PET and HMLS, including MI linked with RSF, enables us to perform the appropriate survival prediction. We also showed imaging information extracted from PET outperformed the usage of the fused images in the prediction of RFS.
Mohammad R. Salmanpour, Ghasem Hajianfar, Mahdi Hosseinzadeh, Seyed Masoud Rezaeijo, Mohammad Mehdi Hosseini, Ehsanhosein Kalatehjari, Ali Harimi, Arman Rahmim
Deep Learning and Radiomics Based PET/CT Image Feature Extraction from Auto Segmented Tumor Volumes for Recurrence-Free Survival Prediction in Oropharyngeal Cancer Patients
Abstract
Aim: The development and evaluation of deep learning (DL) and radiomics based models for recurrence-free survival (RFS) prediction in oropharyngeal squamous cell carcinoma (OPSCC) patients based on clinical features, positron emission tomography (PET) and computed tomography (CT) scans and GTV (Gross Tumor Volume) contours of primary tumors and pathological lymph nodes.
Methods: A DL auto-segmentation algorithm generated the GTV contours (task 1) that were used for imaging biomarkers (IBMs) extraction and as input for the DL model. Multivariable cox regression analysis was used to develop radiomics models based on clinical and IBMs features. Clinical features with a significant correlation with the endpoint in a univariable analysis were selected. The most promising IBMs were selected by forward selection in 1000 times bootstrap resampling in five-fold cross validation. To optimize the DL models, different combinations of clinical features, PET/CT imaging, GTV contours, the selected radiomics features and the radiomics model predictions were used as input. The combination with the best average performance in five-fold cross validation was taken as the final input for the DL model. The final prediction in the test set, was an ensemble average of the predictions from the five models for the different folds.
Results: The average C-index in the five-fold cross validation of the radiomics model and the DL model were 0.7069 and 0.7575, respectively. The radiomics and final DL models showed C-indexes of 0.6683 and 0.6455, respectively in the test set.
Conclusion: The radiomics model for recurrence free survival prediction based on clinical, GTV and CT image features showed the best predictive performance in the test set with a C-index of 0.6683.
Baoqiang Ma, Yan Li, Hung Chu, Wei Tang, Luis Ricardo De la O Arévalo, Jiapan Guo, Peter van Ooijen, Stefan Both, Johannes Albertus Langendijk, Lisanne V. van Dijk, Nanna Maria Sijtsema
Backmatter
Metadata
Title
Head and Neck Tumor Segmentation and Outcome Prediction
Editors
Vincent Andrearczyk
Valentin Oreiller
Mathieu Hatt
Adrien Depeursinge
Copyright Year
2023
Electronic ISBN
978-3-031-27420-6
Print ISBN
978-3-031-27419-0
DOI
https://doi.org/10.1007/978-3-031-27420-6

Premium Partner