TY - JOUR
T1 - Fusion of colour contrasted images for early detection of oesophageal squamous cell dysplasia from endoscopic videos in real time
AU - Gao, Xiaohong
AU - Taylor, Stephen
AU - Pang, Wei
AU - Hui, Rui
AU - Lu, Xin
AU - Braden, Barbara
AU - Allan, Philip
AU - Ambrose, Tim
AU - Arancibia-Cárcamo, Carolina
AU - Bailey, Adam
AU - Barnes, Ellie
AU - Bird-Lieberman, Elizabeth
AU - Bornschein, Jan
AU - Brain, Oliver
AU - Collier, Jane
AU - Culver, Emma
AU - East, James
AU - Geremia, Alessandra
AU - George, Bruce
AU - Howarth, Lucy
AU - Jones, Kelsey
AU - Klenerman, Paul
AU - Leedham, Simon
AU - Palmer, Rebecca
AU - Powrie, Fiona
AU - Rodrigues, Astor
AU - Satsangi, Jack
AU - Simmons, Alison
AU - Travis, Simon
AU - Uhlig, Holm
AU - Walsh, Alissa
N1 - Funding Information:
This work was supported by Cancer Research UK [C ref./A 29021] and by the Cancer Research UK (CR-UK) grant number C5255/A18085, through the Cancer Research UK Oxford Centre, and the Oxford Biomedical Research Centre. Their financial support is gratefully acknowledged.
Publisher Copyright:
© 2022 The Author(s)
PY - 2023/4
Y1 - 2023/4
N2 - Standard white light (WL) endoscopy often misses precancerous oesophageal changes due to their only subtle differences to the surrounding normal mucosa. While deep learning (DL) based decision support systems benefit to a large extent, they face two challenges, which are limited annotated data sets and insufficient generalisation. This paper aims to fuse a DL system with human perception by exploiting computational enhancement of colour contrast. Instead of employing conventional data augmentation techniques by alternating RGB values of an image, this study employs a human colour appearance model, CIECAM, to enhance the colours of an image. When testing on a frame of endoscopic videos, the developed system firstly generates its contrast-enhanced image, then processes both original and enhanced images one after another to create initial segmentation masks. Finally, fusion takes place on the assembled list of masks obtained from both images to determine the finishing bounding boxes, segments and class labels that are rendered on the original video frame, through the application of non-maxima suppression technique (NMS). This deep learning system is built upon real-time instance segmentation network Yolact. In comparison with the same system without fusion, the sensitivity and specificity for detecting early stage of oesophagus cancer, i.e. low-grade dysplasia (LGD) increased from 75% and 88% to 83% and 97%, respectively. The video processing/play back speed is 33.46 frames per second. The main contribution includes alleviation of data source dependency of existing deep learning systems and the fusion of human perception for data augmentation.
AB - Standard white light (WL) endoscopy often misses precancerous oesophageal changes due to their only subtle differences to the surrounding normal mucosa. While deep learning (DL) based decision support systems benefit to a large extent, they face two challenges, which are limited annotated data sets and insufficient generalisation. This paper aims to fuse a DL system with human perception by exploiting computational enhancement of colour contrast. Instead of employing conventional data augmentation techniques by alternating RGB values of an image, this study employs a human colour appearance model, CIECAM, to enhance the colours of an image. When testing on a frame of endoscopic videos, the developed system firstly generates its contrast-enhanced image, then processes both original and enhanced images one after another to create initial segmentation masks. Finally, fusion takes place on the assembled list of masks obtained from both images to determine the finishing bounding boxes, segments and class labels that are rendered on the original video frame, through the application of non-maxima suppression technique (NMS). This deep learning system is built upon real-time instance segmentation network Yolact. In comparison with the same system without fusion, the sensitivity and specificity for detecting early stage of oesophagus cancer, i.e. low-grade dysplasia (LGD) increased from 75% and 88% to 83% and 97%, respectively. The video processing/play back speed is 33.46 frames per second. The main contribution includes alleviation of data source dependency of existing deep learning systems and the fusion of human perception for data augmentation.
KW - Colour contrast enhancement
KW - Deep machine learning
KW - Early squamous cell cancer detection
KW - Endoscopic treatment
KW - gastrointestinal endoscopy
KW - Oesophagus cancer
KW - Surveillance
UR - http://www.scopus.com/inward/record.url?scp=85145649406&partnerID=8YFLogxK
U2 - 10.1016/j.inffus.2022.11.023
DO - 10.1016/j.inffus.2022.11.023
M3 - Article
SN - 1566-2535
VL - 92
SP - 64
EP - 79
JO - Information Fusion
JF - Information Fusion
ER -