Progressive Transformer-Based Generation of Radiology Reports
About
Inspired by Curriculum Learning, we propose a consecutive (i.e., image-to-text-to-text) generation framework where we divide the problem of radiology report generation into two steps. Contrary to generating the full radiology report from the image at once, the model generates global concepts from the image in the first step and then reforms them into finer and coherent texts using a transformer architecture. We follow the transformer-based sequence-to-sequence paradigm at each step. We improve upon the state-of-the-art on two benchmark datasets.
Farhad Nooralahzadeh, Nicolas Perez Gonzalez, Thomas Frauenfelder, Koji Fujimoto, Michael Krauthammer• 2021
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Radiology Report Generation | MIMIC-CXR (test) | BLEU-40.107 | 121 | |
| Radiology Report Generation | IU-Xray (test) | ROUGE-L0.39 | 55 | |
| Medical Report Generation | MIMIC-CXR | BLEU-40.107 | 43 | |
| Medical Report Generation | MIMIC-CXR (test) | ROUGE-L0.272 | 39 | |
| Radiology Report Generation | MIMIC-CXR | ROUGE-L27.2 | 32 | |
| Chest X-ray Report Generation | MIMIC-CXR | Precision24 | 8 |
Showing 6 of 6 rows