A Self-Guided Framework for Radiology Report Generation


Jun Li1,2
Shibo Li1
Ying Hu1
Huiren Tao3

1Shenzhen Institute of Advanced Technology
2University of Chinese Academy of Sciences
3Shenzhen University General Hospital

Code [GitHub]

Paper [arXiv]


Abstract

Automatic radiology report generation is essential to computer-aided diagnosis. Through the success of image captioning, medical report generation has been achievable. However, the lack of annotated disease labels is still the bottleneck of this area. In addition, the image-text data bias problem and complex sentences make it more difficult to generate accurate reports. To address these gaps, we pre-sent a self-guided framework (SGF), a suite of unsupervised and supervised deep learning methods to mimic the process of human learning and writing. In detail, our framework obtains the domain knowledge from medical reports with-out extra disease labels and guides itself to extract fined-grain visual features as-sociated with the text. Moreover, SGF successfully improves the accuracy and length of medical report generation by incorporating a similarity comparison mechanism that imitates the process of human self-improvement through compar-ative practice. Extensive experiments demonstrate the utility of our SGF in the majority of cases, showing its superior performance over state-of-the-art meth-ods. Our results highlight the capacity of the proposed framework to distinguish fined-grained visual details between words and verify its advantage in generating medical reports.



Architecture

Our proposed self-guided framework includes Knowledge Distiller (KD), Knowledge Matched Visual Extractor (KMVE), and Report Generator (RG).



Results

Performance Comparison

In Table 1, our results were compared to other state-of the-art methods.

Ablation study

Table 2 shows the quantitative results of our proposed methods.

Visualization and Examples