TR7Net: A Hybrid Transformer-CNN Framework for Endoscopic Image Segmentation with Validation on Spinal Surgery
Authors:
Shao-Chi Pao, Liuyi Yang, Lin Lin, Fengcheng Mei, Xiaoxing Yang, and Bingding Huang
Conference:
ICIC 2025 Posters, Ningbo, China, July 26-29, 2025
Pages:
1219-1230
Keywords:
Spinal endoscopic surgery, Medical image segmentation, Deep learning, TransUNet, RSU7
Abstract
In endoscopic surgery, accurate segmentation of medical images is indispensable for surgical planning, navigation, and real-time guidance. However, existing segmentation methods often fall short in capturing complex anatomical details and long-range contextual information, which are critical for precise segmentation. To address these challenges, we introduce TR7Net TransUNet-RSU7-Network , a novel hybrid deep learning framework that integrates the strengths of TransUNet and RSU7 architectures. TransUNet excels in global context modeling through its transformer encoder-decoder structure, while RSU7 is renowned for its robust feature extraction capabilities, particularly in handling intricate image features. TR7Net synergizes these two architectures to achieve superior segmentation performance. Extensive experiments on spinal endoscopic datasets demonstrate that TR7Net outperforms both TransUNet and nnUNet regarding segmentation accuracy and robustness in surgical scenarios. This work presents a significant advancement in medical image segmentation for spinal endoscopic surgery, offering a more precise and reliable solution for surgical assistance.
BibTeX Citation:
@inproceedings{ICIC2025,
author = {Shao-Chi Pao, Liuyi Yang, Lin Lin, Fengcheng Mei, Xiaoxing Yang, and Bingding Huang},
title = {TR7Net: A Hybrid Transformer-CNN Framework for Endoscopic Image Segmentation with Validation on Spinal Surgery},
booktitle = {Proceedings of the 21st International Conference on Intelligent Computing (ICIC 2025)},
month = {July},
date = {26-29},
year = {2025},
address = {Ningbo, China},
pages = {1219-1230},
}