Exploring the Design Space of 3D MLLMs for CT
Report Generation
Mohammed Baharoon1,2∗ , Jun Ma1,7∗ , Congyu Fang1,3,5 , Augustin Toma1,4 ,
†
and Bo Wang1,3,5,6,7
1
Vector Institute for Artificial Intelligence, Toronto, Canada
arXiv:2506.21535v1 [eess.IV] 26 Jun 2025
2
Department of Biomedical Informatics, Harvard Medical School, Boston,
Massachusetts
3
Peter Munk Cardiac Centre, University Health Network, Toronto, Canada
4
Medical Biophysics, University of Toronto, Toronto, Canada
5
Department of Computer Science, University of Toronto, Toronto, Canada
6
Department of Laboratory Medicine and Pathobiology, University of Toronto,
Toronto, Canada
7
AI Hub, University Health Network, Toronto, Canada
* Equal contribution
†
Corresponding Author
Abstract. Multimodal Large Language Models (MLLMs) have emerged
as a promising way to automate Radiology Report Generation (RRG). In
this work, we systematically investigate the design space of 3D MLLMs,
including visual input representation, projectors, Large Language Mod-
els (LLMs), and fine-tuning techniques for 3D CT report generation. We
also introduce two knowledge-based report augmentation methods that
improve performance on the GREEN score by up to 10%, achieving the
2nd place on the MICCAI 2024 AMOS-MM challenge. Our results on the
1,687 cases from the AMOS-MM dataset show that RRG is largely in-
dependent of the size of LLM under the same training protocol. We also
show that larger volume size does not always improve performance if the
original ViT was pre-trained on a smaller volume size. Lastly, we show
that using a segmentation mask along with the CT volume improves per-
formance. The code is publicly available at https://github.com/bowang-
lab/AMOS-MM-Solution.
1 Introduction
Computed Tomography (CT) is a cornerstone of modern diagnostic imaging,
offering detailed insights into internal anatomical structures and playing a critical
role in diagnosing a wide range of diseases [7]. However, the rapid increase in the
need of CT examinations presents a significant challenge for radiologists, who
must interpret complex 3D volumetric data and generate comprehensive reports
under tight time constraints [4]. This growing demand places immense pressure
on healthcare systems, often leading to delays in diagnosis and treatment, which
can adversely affect patient outcomes [24].
2 Authors Suppressed Due to Excessive Length
To address this challenge, there has been growing interest in developing au-
tomated systems using Multimodal Large Language Models (MLLMs) for radi-
ology report generation [31,3,2,28,8], leveraging their ability to process medical
images with advanced natural language generation. A representative example is
the Large Language and Vision Assistant (LLaVA) [19], where a vision encoder
processes input images, and a projector transforms the encoded features into the
language embedding space. These projected embeddings are concatenated with
natural language instructions and fed into a language decoder to generate text
responses conditioned on both the image and instruction embeddings. LLaVA
has been extensively explored for the generation of radiology reports from 2D im-
ages [3,13,14], with adaptations for the medical domain. For example, MAIRA-2
uses the LLaVA framework for the generation of X-ray reports with localized
findings, resulting in more grounded reports [3].
Recent research has expanded to MLLMs for 3D medical images, which offer
richer spatial information but introduce computational challenges such as han-
dling volumetric data and managing the high token count [8,28,2,31]. Works,
such as M3D [2] and RadFM [28], adopt the LLaVA framework by using a
vision-to-language embedding projector. To address the high dimensionality of
3D image embeddings, M3D introduces a spatial pooling layer, while RadFM
employs a perceiver module [28].
In this study, we systematically investigate the design space of 3D MLLMs
for radiology report generation from CT scans. Our work explores key architec-
tural choices, including visual representation, projectors, LLMs, and fine-tuning
strategies. We also introduce heuristic report augmentation methods to improve
the completeness of generated reports. The main contributions are summarized
as follows:
• Decoupled architecture design: we decompose MLLMs into plug-and-
play modules, allowing for comprehensive studies to obtain better insights
on how these components contribute to performance.
• Knowledge-based argumentation: we introduce two knowledge-based re-
port augmentation methods to enhance report completeness, increasing the
performance by over 10%, from 0.366 to 0.470, on the GREEN metric.
• State-of-the-art performance: our solution achieved the second place in
the hidden test set of the MICCAI 2024 AMOS-MM challenge.
• Modular implementation: all methods are implemented and open-sourced
in a unified framework, ensuring reproducibility, fair comparisons, and easy
integration of new methods.
2 Methods
In this work, we focus on MLLMs that follow the LLaVA architecture because
of their simple yet effective and popular design [19], which includes an image
encoder, a projector, and an LLM. Specifically, a Vision Transformer (ViT) [5] is
used to extract visual embeddings from an image, followed by a projector to map
these embeddings from the image space to the LLM input space, which are then
Exploring the Design Space of 3D MLLMs for CT Report Generation 3
passed to the LLM as input, along with the query prompt. Next, in Section 2.2,
and Section 2.3 we explore different design choices for each element in the MLLM.
Lastly, in Section 2.4, we introduce our knowledge-based report augmentation
methods, which help to ensure that the final reports are comprehensive.
Fig. 1. The design space of 3D MLLMs. We explore different choices in the selection
of the visual representation, projectors, LLMs, and fine-tuning methods.
2.1 Visual Representations of 3D Inputs
The major challenge to address in representing 3D volumes for ViTs is the large
number of tokens, which leads to a significant computational burden. Differ-
ent techniques have been developed to reduce token count [17,2,28]. AnyRes-
olution [18] has been proposed to embed large image sizes, which divides the
high-resolution input into multiple smaller crops, followed by concatenating and
compressing their embeddings. We adopt this technique for processing larger 3D
CT volumes.
2.2 Projector Variants
A naive MLP projector is used as the baseline to project the visual embeddings
to the same dimension of LLM input tokens. Spatial pooling perceiver (SPP)
projector is the projector proposed in M3D [2], which reduces the number of
tokens while maintaining the 3D spatial structure. TokenPacker [15] reduce the
number of tokens by interpolating visual features to low resolution. We extended
TokenPacker to take in 3D inputs with one image size by expanding the depth
dimension.
4 Authors Suppressed Due to Excessive Length
2.3 Large Language Models and Fine-tuing Methods
We experimented with the instruction-tuned version of Llama 3.1 8B [6], Phi3-
mini [1], Phi3-medium [1], Qwen 2.5 3B [29], Gemma 2B [23], Mistral 7B [11], and
M3D pre-trained LLMs [2]. In addition to freezing the LLMs, we evaluated dif-
ferent fine-tuning techniques: parameter-efficient fine-tuning (PEFT) techniques
like LoRA [9] and DoRA [20], and full fine-tuning.
2.4 Knowledge-base Report Augmentation
We further explore ways to augment the generated reports to ensure complete-
ness, using two introduced methods: Binary-based Questioning (BQ) and naive
normality (NN) augmentation. Fig 2 shows a graphical representations of the
two methods.
Binary-based Questioning (BQ). In the first part of our report aug-
mentation method, we make additional binary-based question inferencing for
common findings on the three regions. Specifically, we first turn reports into
triplets with the format {entity, position, exist} following [27]. For example, the
finding "A nodular low-density lesion is observed in the right lobe of the liver"
becomes {"low-density lesion", "liver", true}. Then, we reformat the triplet into
a question in the format "Is there {entity} in the {position}", with the "exist"
being the ground truth. If no entity or position is detected, the question format
becomes "Is the {position} normal?" or "Can you observe {entity} in this CT
scan?", respectively.
The transformation into triplets was done by prompting GPT o3-mini. How-
ever, this may result in a lot of variations for the same finding statement (for
example, {enlargement, lymph nodes in the retroperitoneum} and {enlargement
of lymph nodes, retroperitoneum}). We then prompt GPT o3-mini again to go
through all these variations and design a mapping that transforms them all into
one common triplet.
We train a model, which we will call the triplet model, to predict the {exist}
variable for each triplet, where triplets are constructed from findings in the
corresponding report. The triplet model is based on Phi3-mini with an M3D
projector, trained for 200 epochs with the same hyperparameters described in
Section 3.2. Moreover, we collect the most common triplets and include them as
questions in all examples throughout training. For these common triplets, if it,
or any of its variations, is already in that CT’s findings, we use the associated
{exist} variable, otherwise, {exist} is set to false. This ensures that the model is
always optimized for answering these common conditions. At inference time, we
first use certain pre-defined keywords to ensure that the finding is not already
mentioned in the generated report. Then, we prompt the model with the common
triplets and map its binary answer to pre-defined findings. For example, an
answer of "True" for the question "Are there nodules in the lung?" would be
mapped to "Nodules are seen in the lungs." and "False" is instead mapped to
"No nodules are seen in the lungs." These questions were designed to be general
Exploring the Design Space of 3D MLLMs for CT Report Generation 5
Fig. 2. Knowledge-base report augmentation. After the MLLM generates a report, we
apply two additional methods to augment the report, which are Binary-based Ques-
tioning (BQ) and naive normality. For BQ, we train a triplet model to answer questions
about common conditions and use its answers to append additional findings to the gen-
erated report. To train the triplet model, we first prepare a set of question-answer pairs
by prompting GPT o3-mini to generate triplets of the entity, location, and existence (in
the format {entity, location, exsist}) using findings from the reference reports. These
triplets are then reformatted to questions, which are used to prompt the triplet model
to generate binary True/False answers based on the CT volume. The value of "exist"
is used as ground truth. At inference time, we prompt our triplet model with common
triplets, which correspond to common findings, and based on its binary answer we
append a positive or negative finding to the report. For naive normality, we append
normal findings for organs that are not mentioned in the generated report after the
BQ step. These two methods go hand-in-hand to ensure that our reports are complete
and do not miss any common findings.
to improve performance on the metric, but this method can be adapted to specific
triplets.
Naive Normality (NN) augmentation. In this method, we predefine a
list of organs and conditions along with their associated normality finding for
each of the three regions based on common findings from the dataset. Using
this list, we then scan through the generated report and add a normality finding
to organs or conditions not mentioned in the generated report. For example,
in a chest report, we can scan through the generated report to find the word
"heart." If it is not mentioned, we add the finding "The heart size and shape
is normal and within limits. The heart is normal." to the report. Moreover, in
normal findings that require two identifiers, like "No pleural effusion is seen in
6 Authors Suppressed Due to Excessive Length
both pleural cavities or bilateral pleural cavities," we look for both the words
"pleural effusion" and "pleural cavities" being in the same sentence.
3 Experiments and Results
3.1 Dataset
All the experiments are conducted using the MICCAI24 AMOS-MM challenge8 ,
originally from the AMOS dataset [10], which includes 1287 and 400 cases for
training and validation. We used the validation set as our test set because the
test set reports were not released. Each case contains one CT scan and the
findings and impressions sections of the associated report for any or all the
three regions: chest, abdomen, and pelvis. Chest findings are avaliable in 30.3%
of cases, abdomen for 99.8%, and pelvis for 86.9%. We follow the challenge
setting and only use the findings as model output. The dataset was gathered from
Longgang District Central Hospital and Longgang District People’s Hospital in
Shenzhen, China.
3.2 Implementation details
A 3D ViT pre-trained on Radiopedia, from M3D [2], was used as the image
encoder with a patch size of 4 × 162 and a volume size of 32 × 2562 . All MLLMs
were trained for 150 epochs using a batch size of 4, with a final learning rate of
5e−5 and a cosine scheduler. We used a simple one-sentence prompt instructing
the LLM to describe the findings in the CT scan. The Hugging Face Transformers
[26] framework was used for all LLM training and inferencing.
For processing larger CT volumes using AnyResolution [18], the volume is
divided into crops that are then embedded and concatenated before being passed
to the LLM. This allows for handling high-resolution data with ViTs that are
pre-trained on lower resolutions, while preserving intricate visual details. We
use this method to process CT volumes of dimension 64 × 5122 using a ViT
pre-trained on CTs of size 32 × 2562 . In addition to the official metric in the
competition, GREEN [21], we also computed the RaTEScore [30], and commonly
used text similarity metrics such as BLEU, ROUGE, and METEOR [22,16,12].
3.3 Results
We report medical report generation results for different LLMs, projectors, and
fine-tuning methods. We also explore different settings such as volume size, pre-
dicting impressions along with findings, and adding segmentation masks. We
use the GREEN score [21] as our base evaluation metric, but we also report
RaTEScore [30] and text similarity metrics such as BLEU [22], ROGUE [16]
and METEOR [12].
8
MICCAI 2024 AMOS-MM challenge: https://www.codabench.org/competitions/
3137/
Exploring the Design Space of 3D MLLMs for CT Report Generation 7
Table 1. Effects of LLM, projector, and fine-tuning method on model performance.
Clinical Metrics NLP Metrics
LLM Projector FT Method GREEN RaTEScore BLEU ROUGE METEOR
Phi-3 mini 4B 0.366 0.573 0.272 0.384 0.357
Phi-3 medium 14B 0.370 0.572 0.269 0.382 0.356
Gemma 2B 0.359 0.572 0.274 0.392 0.365
Llama3.1 8B M3D Frozen 0.359 0.570 0.264 0.379 0.355
Qwen2.5 3B 0.341 0.554 0.252 0.378 0.342
Mistralv0.3 7B 0.353 0.569 0.258 0.374 0.346
M3D Phi-3 0.365 0.571 0.276 0.394 0.360
M3D 0.366 0.573 0.272 0.384 0.357
Phi-3 mini MLP Frozen 0.346 0.563 0.268 0.386 0.359
TP 0.343 0.551 0.259 0.372 0.349
Frozen 0.366 0.573 0.272 0.384 0.357
LoRA 0.336 0.549 0.253 0.361 0.343
Phi-3 mini M3D
DoRA 0.321 0.546 0.251 0.360 0.343
FT 0.271 0.528 0.232 0.341 0.323
LLMs. First, we try multiple LLMs of different sizes ranging from 2B to 14B
in size to gauge how much the LLM matters for this task. We show our findings
in Table 1. Generally, we notice that the task is LLM-independent, for the LLMs
we tried, and using a slightly larger LLM or better LLM does not significantly
increase performance. However, we also notice that Qwen 2.5 performs worse
than all the other models, which could be a result of its training data.
Projector Variants. We also benchmark different vision-to-text embedding
projectors. Our results are shown in Table 1. We notice that the M3D projector
[2] outperforms both TokenPacker [15] and the naive MLP projector. This is
likely due to the projector’s ability to preserve 3D spatial information, compared
to the other methods, which confirms the original work’s findings [2].
Fine-tuning Methods. Moreover, we investigate the effects of using PEFT
techniques like LoRA [9] and DoRA [20], compared to full fine-tuning or keeping
the LLM frozen. Table 1 shows our results. We notice that, generally, increasing
the number of tunable parameters in the LLM correlates negatively with perfor-
mance, with a frozen LLM performing the best. This could be due to the task’s
sensitivity to overfitting or the relatively small scale of our dataset.
Visual Input Representations. We experiment with increasing the image
size and using the AnyResolution method [18]. For AnyResolution, we use a CT
volume of size 64 × 5122 , and process crops of size 32 × 2562 , resulting in 8 non-
overlapping crops being processed in total, per forward pass. We also pass in the
original volume resized to 32 × 2562 , following the original implementation [18].
Table 2 shows our results. Our results show that increasing the image resolution
decreases the score across all metrics, on top of the increased computational
burden. Going from the 32 × 2562 image size to 32 × 5122 increases the number
8 Authors Suppressed Due to Excessive Length
of tokens for the ViT 4 times. This could be due to multiple factors, including
the LLMs inability to handle longer contexts and unreliable evaluation metrics.
Table 2. Effect of increasing image resolution. Generally, we observe that increasing
the image resolution decreases performance when the ViT is pre-trained on 32 × 2562 .
Under similar compute, using AnyRes performs better than not.
Clinical Metrics NLP Metrics
Vol. Size AnyRes GREEN RaTEScore BLEU ROUGE METEOR
32 × 2562 F 0.366 0.573 0.272 0.384 0.357
32 × 5122 F 0.328 0.551 0.250 0.370 0.339
64 × 5122 T 0.344 0.560 0.256 0.379 0.344
Additional Experiments. We further experiment with several different train-
ing scenarios, such as predicting impressions along with findings and feeding a
segmentation mask generated by TotalSegmentor [25], along with the CT vol-
ume, where we embed the volume and organ segmentation mask separately and
concatenate their embeddings before being passed to the LLM. The results are
shown in Table 3. Predicting impressions slightly degrades performance. How-
ever, using the segmentation mask along with the volume slightly improves per-
formance on both GREEN and RaTEScore, which could be because it’s easier
to identify abnormalities and organs with the help of the masks.
Table 3. Additional experiments. Predicting the impressions with findings slightly
degrades performance, and using the segmentation mask improves performance. Only
the findings were used in the evaluation.
Clinical Metrics NLP Metrics
Method GREEN RaTEScore BLEU ROUGE METEOR
Baseline 0.366 0.573 0.272 0.384 0.357
With Impressions 0.356 0.571 0.273 0.388 0.366
With Segmentation 0.372 0.581 0.271 0.392 0.357
Knowledge-base Report Augmentation. Lastly, we report the results
for our introduced knowledge-base report augmentation methods. Table 4 shows
that both of our methods significantly increase the GREEN score, with naive
normality increasing the average score by 8%. We suspect that the dispropor-
tionately large increase in the pelvis GREEN is because the reports contain more
normal findings. An important observation is that GREEN does not always cor-
relate with the other metrics. Naive normality decreased the score text-similarity
metrics and did not have an effect on RaTEScore, which could be indicative of
the metric’s robustness against these types of tricks.
Exploring the Design Space of 3D MLLMs for CT Report Generation 9
Table 4. Knowledge-base report augmentation. Both methods significantly increase
performance for GREEN. However, naive normality does not affect RaTEScore and
decreases text-similarity metrics. Each row builds on the previous one: BQ is added to
the baseline, and Naive Normality (NN) is applied on top of BQ. We report a P-value
< 0.05 across all regions for GREEN and RaTEScore between baseline and the final
reports (BQ + NN).
GREEN RaTEScore BLEU ROUGE METEOR
Method Chest Abdomen Pelvis Avg. Avg. Avg.
Baseline 0.243 0.358 0.499 0.366 0.573 0.272 0.384 0.357
+ BQ 0.260 0.391 0.526 0.392 0.599 0.264 0.410 0.384
+ NN 0.287 0.415 0.708 0.470 0.601 0.207 0.390 0.384
Qualitative Examples of Our Report Augmentation Methods
Ground Truth Generated Augmented
Prostate enlargement is observed. Calcifications are observed
The prostate is slightly enlarged with
in the prostate. The bladder trigone is clear. No abnormalities
smooth contour and multiple calcified
are seen in the bilateral adnexal regions. No nodules are
high-density foci are seen inside. No An enhanced scan shows
observed in the bladder. The uterus is normal, with normal
obvious abnormal enhancement is seen uneven enhancement. The
density and no abnormalities. The intestine is normal. The
during the enhanced scan. The angle bladder wall is smooth
bladder-vesical junction is clear. Bladder is filled. The bilateral
between the bladder and seminal vesicle and no obvious thickening
seminal vesicles are symmetrical, with no abnormal density
is clear. The bladder is not fully filled, is observed. No obvious
inside. The pelvic region is normal, with no soft tissue mass.
and the anterior wall is unevenly enlarged lymph nodes are
The surrounding fat gap is clear. No enlarged lymph nodes are
thickened to about 8mm with no seen in the pelvic cavity. A
seen in the retroperitoneum. An enhanced scan shows uneven
obvious nodules or masses, while the small amount of fluid is
enhancement. The bladder wall is smooth and no obvious
enhancement is uniform. No enlarged found in the pelvic cavity.
thickening is observed. No obvious enlarged lymph nodes
lymph nodes are seen in the pelvic
are seen in the pelvic cavity. A small amount of fluid is
cavity.
found in the pelvic cavity.
Calcifications are observed in the prostate. The bladder trigone
The bladder is filled, with
is clear. No abnormalities are seen in the bilateral adnexal
no thickening of the wall.
regions. No nodules are observed in the bladder. The uterus is
The prostate is not
normal, with normal density and no abnormalities. The
enlarged. No obvious
intestine is normal. The bladder-vesical junction is clear. The
There are scattered spot-like enlargement of lymph
bilateral seminal vesicles are symmetrical, with no abnormal
calcifications in the prostate. nodes in the
density inside. The pelvic region is normal, with no soft tissue
retroperitoneum, and no
mass. The surrounding fat gap is clear. The bladder is filled,
free fluid density is
with no thickening of the wall. The prostate is not enlarged.
observed in the abdominal
No obvious enlargement of lymph nodes in the
and pelvic cavities.
retroperitoneum, and no free fluid density is observed in
the abdominal and pelvic cavities.
No obvious abnormalities in the seminal vesicles. The chest is
The lung texture is clear,
The pulmonary vasculature is clear, symmetrical. Chest bones are normal. No pleural effusion is
with normal distribution
with no abnormal distribution or seen in both pleural cavities or bilateral pleural cavities. No
and no abnormalities. The
infiltrative lesions in the lung infiltrative or space-occupying lesions are seen in the lung
trachea and major
parenchyma. The right subclavian artery parenchyma. The airways are unobstructed. No enlargement is
bronchial branches are
originates from the aortic arch and runs observed in the bilateral pulmonary hila. The Lung fields are
unobstructed. No enlarged
posterior to the esophagus. The clear and normal with no evidence of consolidation. The lung
lymph nodes are seen in
mediastinal window shows no texture is clear, with normal distribution and no
the bilateral hilar and
enlargement of the bilateral hilar lymph abnormalities. The trachea and major bronchial branches
mediastinum. The heart is
nodes, and the trachea and main bronchi are unobstructed. No enlarged lymph nodes are seen in the
normal in size and shape.
are unobstructed. No abnormal findings bilateral hilar and mediastinum. The heart is normal in size
No fluid density is seen in
are noted in the pleura, ribs, or soft and shape. No fluid density is seen in the bilateral pleural
the bilateral pleural
tissues of the chest wall. cavities. No nodules are seen in the lungs. No shadows are
cavities.
seen in the lung.
Fig. 3. Qualitative examples of augmented reports. We show 3 examples of ground
truth, generated, and the augmented reports after applying Binary-based Questioning
and Naive Normality. Findings highlighted in red represents findings that are origi-
nally missed in the generated report, but that are then captured in the augmented
one.Bolded findings in the augmented report represents the original generated report.
All three examples show that our report augmentation methods are able to capture
missed positive findings, while also ensuring completeness by mentioning normal or-
gans explicitly.
10 Authors Suppressed Due to Excessive Length
Qualitative Results. We show in Figure 3 qualitative examples of the aug-
mented reports after applying BQ and NN. Findings highlighted in red represents
findings that are originally missed in the generated report, but are then captured
in the augmented report. In the first report, the enlargement and calcification of
the prostate are missed, but are then captured using additional inferences with
our BQ triplet model. The added findings (For example, "Prostate enlargement
is observed") are directly extracted from the knowledge-base, and this frame-
work can be customizable based on the desired specificity of the findings. The
last example shows an incomplete generated report that does not mention any
observation about regions like the pleura and chest walls. These are then naively
augmented using the NN method. Even though this does not change the inherent
meaning of the report, it ensures that the final report is more explicit.
4 Conclusion
Our study reveals that CT report generation with 3D MLLMs, specifically on
the AMOS-MM dataset, is largely LLM-independent, with models from 2B to
14B parameters showing similar performance. The M3D projector outperformed
simpler methods by preserving 3D spatial structure, and freezing the LLM, com-
pared to using parameter-efficient techniques or full fine-tuning. Moreover, in-
creasing the input image resolution through the AnyResolution method did not
enhance performance, underscoring the need for alignment between the pre-
training and fine-tuning image resolutions. Lastly, our introduced knowledge-
based report augmentation methods significantly boosted the GREEN score by
up to 10%. These insights contribute to a deeper understanding of the architec-
tural choices for optimizing automated medical report generation, providing a
robust foundation for future research in this domain.
Acknowledgement
Our codebase is built upon the M3D [2] GitHub repository. We also gracefully
acknowledge Dr. Fan Bai for his invaluable suggestions and guidance. We also
highly appreciate all the challenge organizers of the MICCAI 2024 AMOS-MM
challenge for allowing us to use this great dataset.
References
1. Marah Abdin, Jyoti Aneja, Hany Awadalla, Ahmed Awadallah, Ammar Ahmad
Awan, et al. Phi-3 technical report: A highly capable language model locally on
your phone. arXiv:2404.14219, 2024.
2. Fan Bai, Yuxin Du, Tiejun Huang, Max Q-H Meng, and Bo Zhao. M3d: Advancing
3d medical image analysis with multi-modal large language models. arXiv preprint
arXiv:2404.00578, 2024.
Exploring the Design Space of 3D MLLMs for CT Report Generation 11
3. Shruthi Bannur, Kenza Bouzid, Daniel C Castro, Anton Schwaighofer, Sam Bond-
Taylor, Maximilian Ilse, Fernando Pérez-García, Valentina Salvatelli, Harshita
Sharma, Felix Meissen, et al. Maira-2: Grounded radiology report generation.
arXiv:2406.04449, 2024.
4. R. J. M. Bruls and R. M. Kwee. Workload for radiologists during on-call hours:
dramatic increase in the past 15 years. Insights into Imaging, 11(1), 2020.
5. Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xi-
aohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg
Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. An image is worth 16x16
words: Transformers for image recognition at scale. In International Conference
on Learning Representations, 2021.
6. Abhimanyu Dubey, Abhinav Jauhri, Abhinav Pandey, Abhishek Kadian, Ahmad
Al-Dahle, Aiesha Letman, Akhil Mathur, Alan Schelten, Amy Yang, Angela Fan,
et al. The llama 3 herd of models. arXiv:2407.21783, 2024.
7. C. J Garvey. Computed tomography in clinical practice. BMJ,
324(7345):1077–1080, 2002.
8. Ibrahim E. Hamamci, Sezgin Er, and Bjoern Menze. Ct2rep: Automated radiology
report generation for 3d medical imaging. In International Conference on Medical
Image Computing and Computer-Assisted Intervention, pages 476–486, 2024.
9. Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean
Wang, Lu Wang, and Weizhu Chen. Lora: Low-rank adaptation of large language
models. arXiv:2106.09685, 2021.
10. Yuanfeng Ji, Haotian Bai, Chongjian Ge, Jie Yang, Ye Zhu, Ruimao Zhang, Zhen
Li, Lingyan Zhanng, Wanling Ma, Xiang Wan, et al. Amos: A large-scale abdominal
multi-organ benchmark for versatile medical image segmentation. Advances in
Neural Information Processing Systems, 35:36722–36732, 2022.
11. Albert Qiaochu Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford,
Devendra Singh Chaplot, Diego de Las Casas, Florian Bressand, Gianna Lengyel,
Guillaume Lample, Lucile Saulnier, L’elio Renard Lavaud, Marie-Anne Lachaux,
Pierre Stock, Teven Le Scao, Thibaut Lavril, Thomas Wang, Timothée Lacroix,
and William El Sayed. Mistral 7b. arXiv preprint arXiv:2310.06825, 2023.
12. Alon Lavie and Abhaya Agarwal. Meteor: an automatic metric for mt evaluation
with high levels of correlation with human judgments. In Proceedings of the Second
Workshop on Statistical Machine Translation, StatMT ’07, page 228–231, USA,
2007. Association for Computational Linguistics.
13. Seowoo Lee, Jiwon Youn, Hyungjin Kim, Mansu Kim, and Soon Ho Yoon. Cxr-
llava: a multimodal large language model for interpreting chest x-ray images. Eu-
ropean Radiology, January 2025.
14. Chunyuan Li, Cliff Wong, Sheng Zhang, Naoto Usuyama, Haotian Liu, Jianwei
Yang, Tristan Naumann, Hoifung Poon, and Jianfeng Gao. Llava-med: Training a
large language-and-vision assistant for biomedicine in one day. Advances in Neural
Information Processing Systems, 36:28541–28564, 2023.
15. Wentong Li, Yuqian Yuan, Jian Liu, Dongqi Tang, Song Wang, Jianke Zhu,
and Lei Zhang. Tokenpacker: Efficient visual projector for multimodal llm.
arXiv:2407.02392, 2024.
16. Chin-Yew Lin. ROUGE: A package for automatic evaluation of summaries. In
Text Summarization Branches Out, pages 74–81, Barcelona, Spain, July 2004. As-
sociation for Computational Linguistics.
17. Che Liu, Zhongwei Wan, Yuqi Wang, Hui Shen, Haozhe Wang, Kangyu Zheng,
Mi Zhang, and Rossella Arcucci. Benchmarking and boosting radiology report
12 Authors Suppressed Due to Excessive Length
generation for 3d high-resolution medical images. arXiv preprint arXiv:2406.07146,
2024.
18. Haotian Liu, Chunyuan Li, Yuheng Li, Bo Li, Yuanhan Zhang, Sheng Shen, and
Yong Jae Lee. Llava-next: Improved reasoning, ocr, and world knowledge, 2024.
19. Haotian Liu, Chunyuan Li, Qingyang Wu, and Yong Jae Lee. Visual instruction
tuning. In Advances in Neural Information Processing Systems, 2023.
20. Shih-Yang Liu, Chien-Yi Wang, Hongxu Yin, Pavlo Molchanov, Yu-Chiang Frank
Wang, Kwang-Ting Cheng, and Min-Hung Chen. Dora: Weight-decomposed low-
rank adaptation. arXiv:2402.09353, 2024.
21. Sophie Ostmeier, Justin Xu, Zhihong Chen, Maya Varma, Louis Blankemeier,
Christian Bluethgen, Arne Md, Michael Moseley, Curtis Langlotz, Akshay Chaud-
hari, et al. Green: Generative radiology report evaluation and error notation. In
Findings of the Association for Computational Linguistics: EMNLP 2024, pages
374–390, 2024.
22. Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. Bleu: a method
for automatic evaluation of machine translation. In Proceedings of the 40th Annual
Meeting on Association for Computational Linguistics, ACL ’02, page 311–318,
USA, 2002. Association for Computational Linguistics.
23. Gemma Team, Morgane Riviere, Shreya Pathak, Pier Giuseppe Sessa, Cassidy
Hardin, Surya Bhupatiraju, Léonard Hussenot, Thomas Mesnard, Bobak Shahriari,
Alexandre Ramé, et al. Gemma 2: Improving open language models at a practical
size. arXiv:2408.00118, 2024.
24. Christopher J. Troupis, Richard A. H. Knight, and Kenneth K. Lau. What is the
appropriate measure of radiology workload: Study or image numbers? Journal of
Medical Imaging and Radiation Oncology, 68(5):530–539, 2024.
25. Jakob Wasserthal, Hanns-Christian Breit, Manfred T. Meyer, Maurice Pradella,
Daniel Hinck, Alexander W. Sauter, Tobias Heye, Daniel T. Boll, Joshy Cyriac,
Shan Yang, Michael Bach, and Martin Segeroth. Totalsegmentator: Robust seg-
mentation of 104 anatomic structures in ct images. Radiology: Artificial Intelli-
gence, 5(5), September 2023.
26. T Wolf. Huggingface’s transformers: State-of-the-art natural language processing.
arXiv:1910.03771, 2019.
27. Chaoyi Wu, Xiaoman Zhang, Ya Zhang, Yanfeng Wang, and Weidi Xie. Med-
klip: Medical knowledge enhanced language-image pre-training in radiology.
arXiv:2301.02228, 2023.
28. Chaoyi Wu, Xiaoman Zhang, Ya Zhang, Yanfeng Wang, and Weidi Xie. Towards
generalist foundation model for radiology by leveraging web-scale 2d&3d medical
data. arXiv preprint arXiv:2308.02463, 2023.
29. An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu,
Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, et al. Qwen2. 5 technical
report. arXiv:2412.15115, 2024.
30. Weike Zhao, Chaoyi Wu, Xiaoman Zhang, Ya Zhang, Yanfeng Wang, and Weidi
Xie. Ratescore: A metric for radiology report generation. arXiv:2406.16845, 2024.
31. Hong-Yu Zhou, Subathra Adithan, Julián Nicolás Acosta, Eric J Topol, and Pranav
Rajpurkar. A generalist learner for multifaceted medical image interpretation.
arXiv preprint arXiv:2405.07988, 2024.