CC BY-NC-ND 4.0 · Endosc Int Open 2020; 08(03): E421-E422
DOI: 10.1055/a-1073-7726
Owner and Copyright © Georg Thieme Verlag KG 2020

Image database, AI and capsule endoscopy; the bets are on

Anastasios Koulaouzidis
1  Endoscopy Unit, Centre for Liver & Digestive Disorders, The Royal Infirmary of Edinburgh, Scotland
Ervin Toth
2  Department of Gastroenterology, Skåne University Hospital, Malmö, Lund University, Sweden
› Author Affiliations
Further Information

Publication History

Publication Date:
21 February 2020 (online)

Capsule has gotten its time certification to become the main endoscopic modality for small bowel inspection. Capsule endoscopy (CE) therefore has become one more technique for the millennial generation that now look at the future with strength, will and determination to go down more disruptive paths [1]. Nevertheless, even for courageous youth, maturing comes with certain “obligations” and societal requirements. Of course, when medical technology is the subject in focus, this translates more or less into diagnostic yield, complication rates, and quality/performance assurance measures [2] [3] [4].

Moreover, in terms of open issues in the field, there are still a few. Perhaps we have little hesitation in counting the indications for CE in a single breath or even providing similar complication rates and pathways for safe “capsuling,” but for those of us who read CE videos in high-volume centers, there are several technological tweaks that we wish the manufacturers would have sorted by now. They range from better image resolution and battery lifetime to various software enhancements and/or hardware advancement(s) [5] [6].

In fact, we kept repeating the cry for the next TIDAL wave in CE revolution, i. e. one that will include Therapy capacity, Integration/Intelligence, Data (not only images) collection, Actuation, and Localization summarizing the main challenges that any – of the many currently available – CE provision systems has to face to increase its market proposal and value over its competitors. The buzz of artificial intelligence (AI) is not new; the very essence of CE reading software, in fact, is one of early AI in action. Tools such as Lewis Score [7], QuickView [8], Suspected Blood Indicator [9] are nothing more than clever snippets of AI integrated in the very early versions of proprietary reading software with main aim to assist, support and/or speed up medical decision process [10].

However, errors and oversights are akin to human nature; when these are associated with erroneous diagnosis, the impact/harm for both patients and healthcare professionals who experience them may be detrimental and associated with loss, emotional hardship, and litigation. On the other hand, striving for improvement is at the heart of all integrated, people-centered health care systems, although that can be a challenging and often slow process. In CE, just as with other imaging specialties such as radiology in which the record of each examination is readily available at any time for review and comparison, the ground is fertile for application of AI to limit findings miss rate, reading time and/or aid lesion characterization and decision-making via computer-aided diagnosis (CAD).

In conventional endoscopy/colonoscopy, systems such as GI Genius, ENDOANGEL and OdinVision are making impact trial after trial but they are all based on/assisted by the human operator in detection of lesions. Nevertheless, CE has the advantage of existing repositories of imaging studies and of course that of “the cloud.” Our deep belief is that the CE should be interpreted on board [11] [12] just like a modern electrocardiogram of the bowels. In 2017, we reported that incorporating machine learning algorithms (MLAs) into CE reading is difficult as large amounts of image annotations are required for training. We had noted back then, presenting KID (a CE interactive database), that existing databases lack graphic annotations of pathologies and cannot be used for advancement of software solutions, maybe the correct term should have been AI [13].

It’s our pleasure to see that Leenhardt et al have been extremely busy recently structuring a framework for a big, national multicentre repository of small bowel CE images [14]. The images have been obtained, for the purpose of clinical care, with PillCamSB3, (Medtronic, Minnesota, United States) from 12 participating centers in France. The study met the approval of the French Data Protection Authority. In total, more than 4,000 CE pooled studies were reviewed and an impressive 6,013 still frames and their respective short video clips were collected. Applying some crowd sourcing, Leenhardt et al used medical students to delineate (annotate) abnormalities in 5,124 frames which was then reviewed by CE experts. This database was called CAD-CAP (Computer-Assisted Diagnosis for CAPsule endoscopy). CAD-CAP is certainly the largest PillCamSB3 database in existence. However, even this database, as the author mentioned, is dwarfed by relevant attempts in ophthalmology and dermatology. The same active group recently presented the use of convolutional neural networks in detection of small bowel angiectasias utilizing this database and therefore proving its research-support potential [15]. Prof Dray’s colonic computed assessment of cleansing (CAC) score was able, based on the ratio of color intensities, to discriminate "adequately cleansed" from "inadequately cleansed" CCE still frames [16]. Furthermore, the very same team is spearheading another major challenge of the times, an attempt for unification of nomenclature of lesions in CE [17].

Essentially, we all aspire that in a few years we will be able to rely on AI to analyze CE videos and present us with its findings. Multiple research and commercial entities across the world are working on respective systems training neural networks. In the meantime, it is certainly beneficial to have efforts and works like those of Leenhardt et al to foster essential multidisciplinary cooperation and make progress in this field. For this reason, we trust that the team will make the images and their annotations available to the wider scientific community and continue with the hard work required to continue populating the database with polyps images and relevant annotations.