SegSTRONG-C

Segmenting Surgical Tools Robustly On Non-adversarial Generated Corruptions

ARCADE Lab, Johns Hopkins University
Descriptive Text

Challenge Description

Accurate segmentation of tools in robot-assisted surgery is a foundational aspect of machine perception, facilitating various downstream tasks, including augmented reality feedback. While existing feed-forward network-based methods exhibit excellent performance in the absence of corruption in test cases, the susceptibility to even minor corruptions can significantly impair the model's performance due to the inherent overfitting nature of such networks. This vulnerability becomes particularly consequential in surgical applications, where high-stakes decisions are commonplace. Prior efforts, such as benchmarking methods on ImageNet-C for image classification, have explored the robustness of models by introducing artificial noise to test images. The CaRTS approach introduces a novel pipeline designed to achieve robust segmentation of robot tools against realistic corruptions. To assess algorithm robustness against non-adversarial corruptions, we expand the dataset in CaRTS to address the challenge of robust robot tool segmentation against non-adversarial corruption. Our goal is to encourage algorithms to exhibit robustness to unforeseen yet plausible complications that may arise during surgery, such as smoke, over-bleeding, and low brightness. The training set comprises 14 mock endoscopic surgery sequences without corruption and corresponding binary segmentation masks where "1" represents robot tools and "0" represents tissue background. In the testing set, there are three sequences with non-adversarial corruptions (smoke, over-bleeding, and low brightness) and digital corruption (ImageNet-C). Additionally, three sequences with an alternative background serve as validation. Participants are challenged to train their algorithms solely on uncorrupted sequences while achieving high performance on corrupted ones for the binary robot tool segmentation task. Successfully achieving this non-adversarial robustness in this benchmark is paramount for the translation of research algorithms into real-world applications. It ensures that these algorithms can navigate and perform effectively in the face of unexpected but reasonable complexities encountered during surgical procedures.

Keywords: Non-adversarial Robustness, Surgical Tool Segmentation, Robotics Surgery, Minimally Invasive Surgery, EndoVis

Task Description

  • Task: Binary segmentation where "1" means tool, "0" means background.
  • Input modality: Images collected via endoscope stereo cameras. They can also be treated as video sequences.
  • Ground truth format:Binary mask for each sequence.

Data Description and Download

  • Total number of cases: We have 6600 cases for training, 3600 cases for validation, and 5400 cases for testing.
  • Important characteristics: We provide 11 sequences of uncorrupted samples for the training, 3 sequences of uncorrupted samples along with alternative background corruption for validation, and two sequences with three kinds of corrupted samples (smoke, over-bleeding, and low-brightness) for the test.
  • Data download: Please contact Hao Ding for data downloading

Metrics and Ranking Methods

  • Metrics: Dice Similarity Coefficient (DSC) and Normalized Surface Distance(NSD) averaged from different tolerances for the robot tool .
  • Performance rank: We rank algorithms with total scores summed up over three test domains - bleeding, smoke, and low brightness. Points are given by the rank for each test domain. Rank from 1 to 5 and get points from 5 to 1 if there are in total of 5 participants. Each Domain will have two ranks based on the DSC and NSD metrics.
  • Submission with missing results: We treat the score of the missing results as zero.

Code base

  • Code base: We provide code base for data loading and baseline evaluation in our github repo

Participants and Test Results

Descriptive Text
Descriptive Text

News

  • 18th Oct 2024: Full dataset released!
  • 12th Oct 2024: Final Results released!
  • 10th Oct 2024: Our challenge is hilighted on MICCAI 2024!
  • 10th Oct 2024: Our challenge is presented on EndoVis MICCAI 2024!
  • 16th July 2024: Baseline results on test set released on the repo and arXiv paper
  • 16th July 2024: The challenge preprint paper on arxiv
  • 8th May 2024: Our code base is updated according to an internal trial.
  • 1st May 2024: We are officially launched!

Organizer

Organization Logo

Organizer: Hao Ding, Tuxun Lu, Yuqian Zhang, Ruixing Liang, Hongchao Shu, Lalithkumar Seenivasan, Yonghao Long (CUHK),Qi Dou(CUHK), Cong Gao (Intuitive), Mathias Unberath

Contact Person: Hao Ding, Mathias Unberath, Cong Gao (Intuitive)

Supported by