tmlr-md-dump / AMIyxTaLgQ /AMIyxTaLgQ.md
RedTachyon's picture
Upload folder using huggingface_hub
5a0f0b3 verified
|
raw
history blame
74.4 kB

Tier: Text-Image Entropy Regularization For Clip-Style Models

Anil Palepu apalepu@mit.edu Harvard-MIT Health Sciences & Technology Andrew L. Beam andrew_beam@hms.harvard.edu Harvard University

Abstract

In this paper, we study the effect of a novel regularization scheme on contrastive languageimage pre-trained (CLIP) models. Our approach is based on the observation that, in many domains, text tokens should only describe a small number of image regions and, likewise, each image region should correspond to only a few text tokens. In CLIP-style models, this implies that text-token embeddings should have high similarity to only a small number of image-patch embeddings for a given image-text pair. We formalize this observation using a novel regularization scheme that penalizes the entropy of the text-token to image-patch similarity scores. We qualitatively and quantitatively demonstrate that the proposed regularization scheme shrinks most of the pairwise text-token and image-patch similarity scores towards zero, thus achieving the desired effect. We demonstrate the promise of our approach in an important medical context, chest x-rays, where this underlying sparsity hypothesis naturally arises. Using our proposed approach, we achieve state of the art (SOTA) average zero-shot performance on the CheXpert and Padchest chest x-ray datasets, outperforming an unregularized version of the model and several recently published self-supervised models.

1 Introduction

Self-supervised vision models that leverage paired text data such as the contrastive language-image pretrained (CLIP) model (Radford et al., 2021; Zhang et al., 2020) have demonstrated very impressive zero-shot classification performance in a variety of domains (Radford et al., 2021; Tiu et al., 2022; Boecking et al., 2022; Palepu & Beam, 2022). Specifically, users can leverage the unified text and image embedding space for zero-shot classification by providing relevant text queries and assessing image embedding similarities (Radford et al., 2021; Tiu et al., 2022; Kumar et al., 2022).

The CLIP architecture consists of a vision encoder, typically a CNN (He et al., 2016) or vision transformer (Dosovitskiy et al., 2020), and a text encoder, typically a text transformer (Vaswani et al., 2017). Each encoder produces a single embedding in the joint embedding space that aims to summarize all of the relevant information in their respective modality. A recent CLIP-style architecture from Boecking et al. (2022), which was built on chest x-ray (CXR) data, allows for a more fine-grained representation of images by projecting the final ResNet block's output to the joint embedding space prior to doing a global average pooling. As a result, this model produces a set of local or patch embeddings which can be indicative of not just if a text and image align, but also roughly where they align. As an example, in a CXR positive for cardiomegaly (an enlarged heart), the patch embeddings near the heart would likely have a higher cosine similarity to the text embedding of "an enlarged heart" than other regions would. In certain domains such as CXRs, it is clear that important visual features tend to be confined to a relatively small portion of the image. For example, cardiomegaly is primarily identified in the lower left portion of the chest, but a CXR captures many anatomical regions beyond this area. At the same time, complex image captions could describe multiple diverse clinical findings which are unlikely to all correspond to the exact same CXR regions. In this work, we propose a method to encode these observations into any CLIP-style model that can produce individual image-patch embeddings and text-token embeddings. To do so, we introduce textimage entropy regularization (TIER), which encourages text-token embeddings and image-patch embeddings to be less 'promiscuous' by regularizing the entropy of a softmaxed distribution of similarity scores. This regularization can be modulated by adjusting two hyperparameters, and because it is based on entropy, it is robust to positional shifts in both the text and the images.

We implement our TIER method leveraging the pre-trained architecture from Boecking et al. (2022), and demonstrate both qualitatively and quantitatively that our regularization method shrinks the text-token and image-patch similarity scores towards zero. We evaluate the resulting model by comparing it to an equivalent unregularized baseline, a fully-supervised baseline, and several state-of-the-art, CLIP-style CXR benchmarks (Tiu et al., 2022; Wang et al., 2022b). We demonstrate that our method results in zero-shot accuracy improvement across a wide range of clinical findings, setting a new state of the art in many instances.

In summary, we make the following contributions:

  • A novel regularization scheme applicable to any CLIP-style model that produces local image and text embeddings. The regularization term shrinks the text-token and image-patch similarity scores to encourage sparser image-text similarities.

  • We establish a new state of the art (SOTA) zero-shot classification AUC on the CheXpert test set, surpassing recently introduced self-supervised models and several previously published fully supervised ones.

  • We also establish a new SOTA for average zero-shot classification AUC on the Padchest dataset, which measures classification performance across many diverse clinical findings.

2 Related Works

Many groups have made efforts to promote more fine-grained alignment of images and text in CLIP-style models (Yao et al., 2021; Li et al., 2022a; Zhong et al., 2022; Wang et al., 2022a; Huang et al., 2021; Li et al., 2022b). Several of these approaches require a separate region proposal/object detection network (Zhong et al., 2022; Li et al., 2022b;a); while typically effective for natural images, these objection detection models have not been equivalently validated in medical domains like CXRs.

Other approaches (Wang et al., 2022a; Huang et al., 2021) aim to modify the contrastive loss in order to better align local representations, but unlike our approach, do not directly aim to induce sparsity. For example, while Huang et al. (2021) and Wang et al. (2022a) both include a local contrastive loss, their approaches potentially allow tokens to be similar to all of the cross-modal tokens. Furthermore, Huang et al.

(2021) was shown to have poor zero-shot performance when evaluated by Tiu et al. (2022), while Wang et al.

(2022a) is not yet publicly available for evaluation.

Unlike the previously described approaches, and like our TIER method, the approach in Yao et al. (2021) does induce sparsity at the token and patch level. However, their approach more aggressively forces sparsity by only considering the maximum similarity text token for each image token and vice versa. Conversely, our approach allows us to flexibly modulate the level of sparsity using two tune-able hyperparameters (which could allow us to mimic the effect of Yao et al. (2021) if set extremely high).

3 Methods 3.1 Data

We utilized the MIMIC-CXR-JPG (Johnson et al., 2019) dataset to train our models and the CheXpert (Irvin et al., 2019) and Padchest (Bustos et al., 2020) datasets to evaluate them. The MIMIC-CXR dataset (Johnson et al., 2019) consists of 377,095 CXR samples from 65,379 different patients. Many patients have multiple radiological studies within the dataset, with a single study often

2_image_0.png

Figure 1: An overview of TIER, our regularized training method. containing both a frontal and lateral CXR view. These CXRs were evaluated by radiologists, who wrote detailed reports on the clinical findings they observed as well as a sentence or two describing their overall impression of the imaging. We extracted these impression sections from the radiology reports to use as the paired text for our image input. We dropped any samples that were missing this impression section, leaving us with a total of 319,446 CXR-impression pairs. We split these MIMIC-CXR image-text pairs into training and validation subsets (with approximately 90% training data) and ensured that no patient was represented in both subsets. For evaluation, we utilized the separate CheXpert (Irvin et al., 2019) dataset with pre-defined validation and test splits which consisted of 234 and 668 CXRs respectively. These subsets of CheXpert have 14 different clinical labels, determined by consensus of 3 and 5 radiologists respectively. We benchmarked our models' thresholded predictions using labels from an additional 3 radiologists available in the CheXpert test set. For the purposes of our evaluation, we only considered the following 5 clinical labels: Cardiomegaly, Edema, Consolidation, Atelectasis, and Pleural Effusion. These labels were the five competition tasks from the CheXpert competition and the most commonly attempted tasks in the literature, making them a natural set for comparison. We also extracted these labels from the MIMIC-CXR dataset, but we only used them when training our fully supervised CNN baseline; our contrastive models did not have any access to these labels. We additionally evaluated our models with the Padchest dataset (Bustos et al., 2020), of which we only considered the subset of 39,053 CXRs that were labeled by radiologists. There were over a hundred different labels present in these CXRs, but we focused on the set of 57 labels that were present with frequency of at least 50 in our selected subset, as was done by Tiu et al. (2022). All images were resized to 224 x 224 pixels with 3 RGB channels. At train time, we performed random data augmentations including random resizing, cropping, affine transformation, and color jitter, while at test time, we simply resized images to 256 x 256 before center cropping to 224 x 224.

3.2 Model Architecture

We based our model on the BioViL architecture (Boecking et al., 2022), which consists of a pre-trained ResNet-50 architecture as the vision encoder and "CXR-BERT-specialized", a transformer, as the text encoder. This model differed from the original CLIP architecture in that it consisted of a radiology-specific text encoder (CXR-BERT-specialized) and was trained with an additional MLM loss, among several other changes (Boecking et al., 2022). This model was also trained using MIMIC-CXR and importantly did not have access to the CheXpert or Padchest datasets, which we used for evaluation. For our purposes, the most critical feature of the BioViL model is that the final ResNet-50 block provides embeddings that correspond to local, connected regions of the input image (see the green path on the top of Figure 1). Thus, in addition to the single global image embedding, for an input image of size 224 × 224 this model also produces a set of 49 embeddings in a 7 × 7 grid, which all share the same joint feature space as the global image embedding. The number of embeddings is a function of the original input size (a larger image input would yield more embeddings) as well as the choice to use the final ResNet block output (using an earlier output would lead to more fine-grained local embeddings). A single multi-layer perceptron with one hidden layer was used to project each local embedding to the joint feature space. We call the output of the ResNet block the patch embeddings as they correspond to regional patches of the image.

The text transformer naturally produces a text token embedding for each input token to the model. We use a single multi-layer perceptron with one hidden layer to project these text token embeddings to the joint feature space. The projected embedding from the first text token, [CLS], is contrasted with the global image embedding as is done with typical contrastive language-image pre-trained models (Radford et al., 2021; Palepu & Beam, 2022; Zhang et al., 2020). For a training batch of image-text pairs (x I, xT ), we use the standard CLIP loss as described in Radford et al. (2021). We add additional penalty terms, described in the following section, to regularize our model beyond this standard CLIP loss. The pseudocode for our method is described in the appendix Fig. 6

3.3 Tier: Text-Image Entropy Regularization Of Image-Patch And Text-Token Similarity Scores

TIER works by first computing a matrix of image-patch and text-token similarities. Specifically, consider an example image-text pair that has I = {I1*, . . . , IP } image-patch embeddings (in our case, P = 49), and has T = {T1, . . . , T*T } text-token embeddings (T can vary for each sample as captions can be different lengths). We compute the image patch-text token similarity matrix S by computing a T ×P matrix of cosine similarities between each image-patch embedding and text-token embedding. The embeddings for each input modality are the outputs of an encoder model that is specific to that input, e.g. a CNN or vision transformer for images and a BERT-style transformer for text. Importantly, we select encoders that provide embeddings at the token level, i.e., image-patch embeddings and text-token embeddings. Row i of S indicates the cosine similarity between a text-token Ti and each image patch in I. The columns of S likewise indicate the cosine similarity between a given image patch Ij and each text token in T.

Recall, the goal of our approach is to shrink the elements of S such that each text token is similar to a relatively small number of image patches. To do this, we introduced an entropy-based penalty term that induces shrinkage on the elements of S. First, we perform a row-wise softmax of S and measure the entropy between a text token Ti and all of the image patches in I, shown below:

H(Ti,I)=j=1Ppjlog(pj){\mathcal{H}}(T_{i},I)=\sum_{j=1}^{|P|}-p_{j}*\log(p_{j}) (1)(1) −pj ∗ log(pj ) (1) where pj is the probability produced by the softmax of the row of S corresponding to Ti. This term will be maximized when each pj is 1P , implying that all of the image patch embeddings have equal similarity to Ti.

Next, we apply the same procedure to the columns of S, applying a column-wise softmax over the text-token similarities to produce probabilities p1 to pT for each image patch Ii and calculating the entropy of these probabilities as follows:

H(T,Ij)=i=1Tpilog(pi){\mathcal{H}}(T,I_{j})=\sum_{i=1}^{|T|}-p_{i}*\log(p_{i})

(2)(2) −pi ∗ log(pi) (2) We average the N ×T image-patch entropies H(Ti, I) and the N ×P text-token entropies H(T, Ij ) to produce an image-patch penalty and text-token penalty for the batch. We control the effects of these penalties on training by weighting them with hyperparameters λt and λp respectively, adding the weighted penalties to the CLIP loss to compute the total loss.

A grid search over the range [0, 0.25] was used to set the hyperparameters (λp = 0.2, λt = 0.1) for our regularized model. Specifically, we trained our contrastive models for just a single epoch on MIMIC-CXR with pairs of λp and λt, and chose the pair that maximized zero-shot AUC on the validation set. These results are available in the appendix Appendix A.3. Both the training procedure and zero-shot classification method are described in later sections.

3.4 Training Details

We begin with the pretrained BioViL architecture and model weights, "CXR-BERT-specialized" (Boecking et al., 2022), which has already been trained with contrastive learning on the MIMIC-CXR dataset. In this original training, only frontal images were used, and they used a masked language model (MLM) loss in addition to the CLIP loss. Starting with this pretrained model, we train two separate CLIP-style models: A regularized model in which λp = 0.2 and λt = 0.1, as well as an unregularized baseline model, in which λp = λt = 0. Despite only minor changes (further training on MIMIC-CXR, inclusion of lateral CXRs, omission of the MLM loss, freezing of early text encoder layers), our unregularized baseline significantly outperformed the publicly available pretrained model from Boecking et al. (2022) as seen in Appendix A.6.

All aspects of model training are identical between our regularized and unregularized models, other than the additional penalty terms. For both models, we freeze the first 8 layers of the BERT encoder, while leaving the rest of the text encoder and vision encoder unfrozen. Each model is trained for 30 epochs using the loss described in the previous section with a learning rate of 0.0001 and batch size of 32.

We also train a fully supervised CNN baseline, which utilizes the same vision encoder as the contrastive models but has a multilayer perceptron with one hidden layer and five outputs. This supervised baseline still uses MIMIC-CXR for training, but instead of text, it is trained with labels using binary-cross entropy loss with a learning rate of 0.0001 and batch size of 32.

3.5 Zero-Shot Classification

We employ a zero-shot classification procedure that leverages our text and image encoders to identify labels of interest in the images. Our method begins with the user selecting Kp positive and Kn negative queries for the label of interest, Q. Positive queries {Qp1*, ..., QpKp } are text descriptions indicative of the presence of that label, while negative queries {Qn1, ..., Q*nKn } are text descriptions indicative of the absence of that label; examples which we used for the five CheXpert labels are detailed in Tab. 8 in the appendix. We pass each positive query through the text encoder, project their [CLS] token embeddings to the joint embedding space, and then average these projected embeddings and re-normalize to a unit norm. We do the same for the negative queries so that we have a single positive Qp and negative Qn query embedding associated with each label that we wish to classify:

Qp=j=1KpQpj/Kpj=1KpQpj/Kp\overline{{{Q_{p}}}}=\frac{\sum_{j=1}^{K_{p}}Q_{p j}/K_{p}}{||\sum_{j=1}^{K_{p}}Q_{p j}/K_{p}||} Qn=j=1KnQnj/Kni=1KnQnj/Kn(3)\overline{Q_{n}}=\frac{\sum_{j=1}^{K_{n}}Q_{nj}/K_{n}}{||\sum_{i=1}^{K_{n}}Q_{nj}/K_{n}||}\tag{3}

For any input image we wish to classify, we use the image encoder to compute its projected global image embedding Eimg (normalized to unit norm) and take the dot product of this global image embedding with both the positive and negative query embeddings Qp and Qn for every label we wish to predict. We subtract these positive and negative cosine similarity scores to get a zero-shot classification score, ZQ, for our label of interest.

ZQ=(EimgQp)(EimgQn)Z_{Q}=(E_{i m g}\cdot{\overline{{Q_{p}}}})-(E_{i m g}\cdot{\overline{{Q_{n}}}}) (4)\left({4}\right) ZQ = (Eimg · Qp) − (Eimg · Qn) (4) Importantly, our zero-shot classification output cannot be interpreted as a probability as its range is actually between [−2, 2]. We are primarily interested in assessing discriminative performance of our zero-shot classifiers, so interpretation as a probability is not necessary; however, if one desired a probability output, they could simply apply a softmax to the positive and negative similarity scores as was done by Tiu et al.

(2022) instead of subtracting these scores.

4 Results 4.1 Visualization Of The Effect Of Regularization

Qualitatively, our regularization method is able to achieve the desired shrinkage between image patches and text tokens. Fig. 2 and Fig. 3 show patch-level zero-shot classification scores (i.e., the score between each image patch and the global [CLS] text token) overlaid on top of two CXRs, one with cardiomegaly and one without. In these heatmaps, red is indicative of a higher zero-shot score, blue is a lower score, while gray is a more neutral score. Important differences between the regularized and unregularized models are apparent when we examine the distribution of blue and red regions of the heatmaps in Fig. 2 and Fig. 3. For the cardiomegaly-positive image (Fig. 2), the regularized model has high score primarily on the lower left side of the patient's chest (which corresponds to lower right side of the image), where their heart is located. Likewise, the regularized model shows low similarity (blue) on the other of their lower chest where one could expect to see some changes in more extreme cases of cardiomegaly. These similarity scores seem rational and are clinically justifiable. On the other hand, while the unregularized model also displays some signal in the clinically relevant regions, it has significantly more extreme similarity scores scattered throughout the image well beyond the heartadjacent regions. We see similar results on an example for pleural effusion detection (Fig. 3), as well as in additional examples presented in the appendix (Fig. 7, Fig. 8, Fig. 9).

4.2 Distribution Of Image-Patch Similarity Scores To Global [Cls] Text Token

To further explore the effect of our regularization method on the image-patch similarities, we utilize a set of 160 positive image-text pairs from MIMIC-CXR. In Fig. 4, we plot the similarity of the projected [CLS] token embeddings to the 49 image-patch embeddings from the corresponding image. In this figure, the image-patch similarities were ranked in descending order before being plotted, with error bars indicating the standard deviation across the 160 samples. We can see that the regularized model on average has significantly lower similarities to the patch embeddings than the unregularized model. To better visualize these differences, we produced Fig. 5, which displays the same information as Fig. 4 except each patch similarity was normalized by dividing the similarity by the sum of all patch similarities in the entire image. In this plot, we can clearly see that the regularized model tends to have a few patches with relatively higher similarities to the [CLS] token embedding, and many with relatively lower similarities; this supports our hypothesis that our regularization scheme shrinks token-level similarity in the model, thus achieving a lower entropy.

4.3 Zero-Shot Classification

Next, we evaluated our zero-shot classification method for both the regularized and unregularized models on the held-out CheXpert test set. Our primary benchmark for these models is the 'CheXzero' model (Tiu et al., 2022), which recently achieved SOTA zero-shot AUC on this task. We use weights from the checkpoint that achieved the highest AUC on the CheXpert validation set. We also evaluate another recent self-supervised model, MedCLIP (Wang et al., 2022b), with the caveat that this model is not strictly zero-shot because the authors utilized clinical labels during their training process. Additionally, we evaluate a fully supervised CNN that uses our vision encoder with an additional classification head. We bootstrap 1000 times, randomly

6_image_0.png

Figure 2: The penalty term induces shrinkage in the image patch-text token similarity scores.

A CXR positive for cardiomegaly, overlaid with heatmaps displaying zero-shot cardiomegaly score for the unregularized (center) and regularized (right) models. Gray corresponds to a neutral (close to zero) zero-shot score, while red is a higher score and blue is a lower score. As can be seen by comparing the middle and right figures, the regularized model focuses more on the relevant regions and less on the periphery. Note: In this instance cardiomegaly is located in the lower right portion of the image.

6_image_1.png

Figure 3: A CXR with pleural effusion, overlayed with heatmaps displaying zero-shot pleural effusion score for the unregularized (center) and regularized (right) models.

6_image_2.png

6_image_3.png

Figure 4: The similarities of each patch to the CLS embedding for a set of 160 MIMIC-CXR, images, sorted in descending order.

Figure 5: The same plot as Fig. 4 in which each similarity is divided by the sum of the total similarity across all patches for that image. 7

Label Regularized TIER Unregularized Basemodel CheXzero MedCLIP Fully Supervised CNN
Average 0.903336 0.89721 0.89349 0.87708 0.88877
Cardiomegaly 0.91714 0.89239 0.88340 0.83911 0.86400
Edema 0.92423 0.90729 0.89424 0.91242 0.92236
Consolidation 0.89712 0.91213 0.91318 0.88653 0.86003
Atelectasis 0.86531 0.85741 0.84299 0.79423 0.85869
Pleural Effusion 0.91290 0.91681 0.93368 0.95313 0.93876

Table 1: Average AUCs for various models on 1000 bootstraps of CheXpert Test. The highest zero-shot AUC is bolded (the three on the left are performing zero-shot classification, in which they have never previously seen any labels in the training set). As seen in the appendix (Tab. 4, Tab. 5, and Tab. 6), all differences between zero-shot models are statistically significant except for Unregularized vs CheXzero for Consolidation.

Though MedCLIP is trained with contrastive learning, it also utilizes labels during training so we do not consider it to be fully zero-shot. sampling the test set with replacement and evaluating the mean AUC performance of each model over these 1000 bootstraps. These results can be seen in Tab. 1, which demonstrates that our regularized model achieves SOTA zero-shot AUC; regularization offers a modest bump in average AUC performance. Excluding "Unregularized vs CheXzero for Consolidation", all pairwise AUC differences between the zero-shot models are statistically significant according to a two sample t-test for the difference of means. These zero-shot models are also competitive against three reference radiologists according to their Matthews's correlation coefficient (MCC) and F1 scores, as seen in the appendix (Fig. 10). We use the Padchest dataset to evaluate a broader set of findings, specifically looking at the 57 findings with n ≥ 50 from the radiologist-labeled subset of Padchest. We constructed positive label queries using the phrase "X is present.", while we constructed negative label queries with the phrase "No X.", replacing X with the label of interest. The notable exception was when we classified "normal" images; in this instance, we used "Abnormal findings." as the negative query. Tab. 2 details the Padchest results for the regularized, unregularized, and CheXzero models. As seen in the appendix (Tab. 9, Tab. 10), our regularized TIER model achieves statistically significant boosts in performance on average as well as for the majority of Padchest findings when compared head-to-head with CheXzero and our unregularized baseline model.

4.4 Zero-Shot Covid-19 Diagnosis

We also evaluate our model COVID-19 detection, which is a diagnosis not present in any of our training data. As a result, our models cannot rely on the actual label itself (i.e., the word cardiomegaly in a text query), and therefore the diagnostic capability of our models on this task can be fully attributed to its ability to recognize the descriptive attributes being queried. Furthermore, discriminating COVID-19 and non-COVID-19 pneumonia from chest imaging is a non-trivial task, with one study reporting just a 74% average accuracy for three radiologists using chest CT for this task (Bai et al., 2020).

We created queries to discriminate COVID-19 and non-COVID-19 pneumonia based on differences mentioned in the literature (Bai et al., 2020; Borghesi & Maroldi, 2020). For the positive COVID-19 query, we used the query "Ground glass opacities and consolidation with peripheral distribution with fine reticular opacity and vascular thickening.", and for the negative COVID-19 query, we used "Pleural effusion present with lymphadenopathy and consolidation with central distribution." (which were described by Bai et al. (2020) as findings more specific to non-COVID-19 pneumonia). We achieved zero-shot AUCs of 0.759, 0.753, and 0.752 with the regularized, unregularized, and CheXzero models respectively on discriminating COVID-19 from non-COVID pneumonia within the COVID-QU-Ex Dataset (Tahir et al., 2021; 2022). This performance indicates that we can leverage our model for difficult multi-class classification tasks by simply providing English descriptions of the class-discriminating features. Furthermore, this procedure can easily extend to other labels, meaning self-supervised vision-language architectures such as these could be leveraged to diagnose novel diseases if their presentation on imaging can be described.

Label Count Regularized AUC Unregularized AUC CheXzero AUC
Average AUC 39053 0.755420 0.742534 0.726306
Number of Evaluations Won (Percent) 57 20 (35.1%) 18 (31.6%) 18 (31.6%)
endotracheal tube 284 0.979606 0.956634 0.98295
pleural effusion 1748 0.942045 0.930292 0.950519
pulmonary edema 87 0.941289 0.945415 0.95646
heart insufficiency 546 0.926097 0.927177 0.917819
pulmonary fibrosis 166 0.951654 0.944147 0.921793
cardiomegaly 3746 0.883692 0.883339 0.890467
vascular redistribution 129 0.877236 0.872019 0.750592
consolidation 364 0.878342 0.849872 0.865175
hilar congestion 601 0.855435 0.850243 0.825707
pulmonary mass 247 0.844107 0.872299 0.842056
cavitation 122 0.857639 0.794295 0.853367
alveolar pattern 1353 0.87631 0.816974 0.763811
calcified pleural thickening 102 0.859651 0.84287 0.850707
lung metastasis 89 0.877375 0.860837 0.827675
emphysema 376 0.717841 0.718377 0.830578
interstitial pattern 1907 0.835144 0.840368 0.816432
costophrenic angle blunting 1683 0.769921 0.808131 0.69029
COPD signs 4823 0.650859 0.652912 0.751217
tuberculosis 59 0.838961 0.843741 0.7978
atelectasis 676 0.781707 0.791507 0.809232
reticular interstitial pattern 72 0.844479 0.867637 0.822429
pneumonia 1780 0.813796 0.796614 0.773941
lobar atelectasis 168 0.808411 0.815725 0.775147
normal 12694 0.776328 0.790588 0.753171
pleural thickening 213 0.784428 0.754608 0.752537
reticulonodular interstitial pattern 51 0.862346 0.838374 0.841414
infiltrates 1456 0.742854 0.735399 0.747836
hypoexpansion 166 0.853423 0.871452 0.794564
hypoexpansion basal 119 0.889652 0.874477 0.8018
humeral fracture 81 0.742305 0.672935 0.749084
pneumothorax 98 0.730643 0.728547 0.777442
multiple nodules 102 0.790815 0.852951 0.716911
hyperinflated lung 197 0.700879 0.667704 0.713202
bronchiectasis 667 0.734643 0.743998 0.690117
adenopathy 136 0.678726 0.73105 0.703924
mediastinal enlargement 106 0.72538 0.666796 0.759299
laminar atelectasis 1378 0.67343 0.687839 0.679276
vertebral compression 126 0.723955 0.734413 0.646448
rib fracture 140 0.689835 0.668069 0.691037
tuberculosis sequelae 185 0.796895 0.773832 0.584302
hilar enlargement 447 0.721779 0.714687 0.678564
tracheal shift 180 0.615827 0.500734 0.634359
mediastinal mass 74 0.709825 0.409473 0.647695
central vascular redistribution 63 0.728932 0.567387 0.354491
vertebral fracture 104 0.791375 0.86009 0.499654
superior mediastinal enlargement 153 0.551017 0.637878 0.596948
vascular hilar enlargement 1428 0.625607 0.60417 0.623934
nodule 736 0.446317 0.507929 0.546737
air trapping 1952 0.580408 0.631534 0.580882
bullas 192 0.744606 0.584846 0.486494
ground glass pattern 123 0.671321 0.661248 0.602802
calcified adenopathy 124 0.673757 0.624151 0.583562
minor fissure thickening 127 0.600411 0.558331 0.77315
unchanged 4036 0.618171 0.633874 0.395541
clavicle fracture 74 0.596974 0.596031 0.607514
pseudonodule 795 0.476977 0.472281 0.557981
end on vessel 63 0.397635 0.485072 0.560626

Table 2: On average over 1000 bootstraps, the regularized model outperforms the unregularized and CheXzero models in fine-grained finding prediction. Zero-shot AUCs for 57 padchest findings. The best AUC for each finding across the three tested models is shown in bold. As shown in Tab. 9 and Tab. 10, all winners are statistically significant except for CheXzero for the rib fracture finding.

5 Discussion & Limitations

In this work, we introduce a regularization method for contrastive language-image pre-trained models which encourages shrinkage of the image-patch and text-token similarities. We demonstrate how our regularization method can benefit zero-shot performance of these models by training a model that achieves SOTA zero-shot classification performance on a broad set of CXR findings. The improvements were robust across a wide range of tasks relative to many strong benchmarks, though in some instances the improvements were modest. Though our work was confined to a medical context, we believe it should be broadly applicable to many other areas where CLIP-style models are used, though these applications were beyond the scope of the present work. We believe our work contributes to a growing literature (Kumar et al., 2022; Mu et al., 2022; Meier et al., 2021) seeking to augment and improve CLIP-style models with inductive biases and domain-specific observations.

References

Harrison X Bai, Ben Hsieh, Zeng Xiong, Kasey Halsey, Ji Whae Choi, Thi My Linh Tran, Ian Pan, Lin-Bo Shi, Dong-Cui Wang, Ji Mei, et al. Performance of radiologists in differentiating covid-19 from non-covid-19 viral pneumonia at chest ct. Radiology, 296(2):E46–E54, 2020. 8 Benedikt Boecking, Naoto Usuyama, Shruthi Bannur, Daniel C Castro, Anton Schwaighofer, Stephanie Hyland, Maria Wetscherek, Tristan Naumann, Aditya Nori, Javier Alvarez-Valle, et al. Making the most of text semantics to improve biomedical vision–language processing. arXiv preprint arXiv:2204.09817, 2022. 1, 2, 4, 5, 16 Andrea Borghesi and Roberto Maroldi. Covid-19 outbreak in italy: experimental chest x-ray scoring system for quantifying and monitoring disease progression. La radiologia medica, 125(5):509–513, 2020. 8 Aurelia Bustos, Antonio Pertusa, Jose-Maria Salinas, and Maria de la Iglesia-Vayá. Padchest: A large chest x-ray image dataset with multi-label annotated reports. Medical image analysis, 66:101797, 2020. 2, 3 Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020. 1 Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016. 1 Shih-Cheng Huang, Liyue Shen, Matthew P Lungren, and Serena Yeung. Gloria: A multimodal globallocal representation learning framework for label-efficient medical image recognition. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pp. 3942–3951, 2021. 2 Jeremy Irvin, Pranav Rajpurkar, Michael Ko, Yifan Yu, Silviana Ciurea-Ilcus, Chris Chute, Henrik Marklund, Behzad Haghgoo, Robyn Ball, Katie Shpanskaya, et al. Chexpert: A large chest radiograph dataset with uncertainty labels and expert comparison. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pp. 590–597, 2019. 2, 3 Alistair EW Johnson, Tom J Pollard, Nathaniel R Greenbaum, Matthew P Lungren, Chih-ying Deng, Yifan Peng, Zhiyong Lu, Roger G Mark, Seth J Berkowitz, and Steven Horng. Mimic-cxr-jpg, a large publicly available database of labeled chest radiographs. arXiv preprint arXiv:1901.07042, 2019. 2 Bhawesh Kumar, Anil Palepu, Rudraksh Tuwani, and Andrew Beam. Towards reliable zero shot classification in self-supervised models with conformal prediction. arXiv preprint arXiv:2210.15805, 2022. 1, 10 Juncheng Li, Xin He, Longhui Wei, Long Qian, Linchao Zhu, Lingxi Xie, Yueting Zhuang, Qi Tian, and Siliang Tang. Fine-grained semantically aligned vision-language pre-training. arXiv preprint arXiv:2208.02515, 2022a. 2 Liunian Harold Li, Pengchuan Zhang, Haotian Zhang, Jianwei Yang, Chunyuan Li, Yiwu Zhong, Lijuan Wang, Lu Yuan, Lei Zhang, Jenq-Neng Hwang, et al. Grounded language-image pre-training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 10965–10975, 2022b.

2 Joshua Meier, Roshan Rao, Robert Verkuil, Jason Liu, Tom Sercu, and Alex Rives. Language models enable zero-shot prediction of the effects of mutations on protein function. Advances in Neural Information Processing Systems, 34:29287–29303, 2021. 10 Norman Mu, Alexander Kirillov, David Wagner, and Saining Xie. Slip: Self-supervision meets languageimage pre-training. In European Conference on Computer Vision, pp. 529–544. Springer, 2022. 10 Anil Palepu and Andrew L Beam. Self-supervision on images and text reduces reliance on visual shortcut features. arXiv preprint arXiv:2206.07155, 2022. 1, 4 Alec Radford, Jong Wook Kim, Chris Hallacy, Aditya Ramesh, Gabriel Goh, Sandhini Agarwal, Girish Sastry, Amanda Askell, Pamela Mishkin, Jack Clark, et al. Learning transferable visual models from natural language supervision. In International Conference on Machine Learning, pp. 8748–8763. PMLR, 2021. 1, 4, 16 Anas M Tahir, Muhammad EH Chowdhury, Amith Khandakar, Tawsifur Rahman, Yazan Qiblawey, Uzair Khurshid, Serkan Kiranyaz, Nabil Ibtehaz, M Sohel Rahman, Somaya Al-Maadeed, et al. Covid-19 infection localization and severity grading from chest x-ray images. Computers in biology and medicine, 139: 105002, 2021. 8 Anas M. Tahir, Muhammad E. H. Chowdhury, Yazan Qiblawey, Amith Khandakar, Tawsifur Rahman, Serkan Kiranyaz, Uzair Khurshid, Nabil Ibtehaz, Sakib Mahmud, and Maymouna Ezeddin. Covid-qu-ex dataset, 2022. URL https://www.kaggle.com/dsv/3122958. 8 Ekin Tiu, Ellie Talius, Pujan Patel, Curtis P Langlotz, Andrew Y Ng, and Pranav Rajpurkar. Expertlevel detection of pathologies from unannotated chest x-ray images via self-supervised learning. Nature Biomedical Engineering, pp. 1–8, 2022. 1, 2, 3, 6 Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia Polosukhin. Attention is all you need. Advances in neural information processing systems, 30, 2017. 1 Fuying Wang, Yuyin Zhou, Shujun Wang, Varut Vardhanabhuti, and Lequan Yu. Multi-granularity crossmodal alignment for generalized medical visual representation learning. arXiv preprint arXiv:2210.06044, 2022a. 2 Zifeng Wang, Zhenbang Wu, Dinesh Agarwal, and Jimeng Sun. Medclip: Contrastive learning from unpaired medical images and text. arXiv preprint arXiv:2210.10163, 2022b. 2, 6 Lewei Yao, Runhui Huang, Lu Hou, Guansong Lu, Minzhe Niu, Hang Xu, Xiaodan Liang, Zhenguo Li, Xin Jiang, and Chunjing Xu. Filip: fine-grained interactive language-image pre-training. arXiv preprint arXiv:2111.07783, 2021. 2 Yuhao Zhang, Hang Jiang, Yasuhide Miura, Christopher D Manning, and Curtis P Langlotz. Contrastive learning of medical visual representations from paired images and text. arXiv preprint arXiv:2010.00747, 2020. 1, 4 Yiwu Zhong, Jianwei Yang, Pengchuan Zhang, Chunyuan Li, Noel Codella, Liunian Harold Li, Luowei Zhou, Xiyang Dai, Lu Yuan, Yin Li, et al. Regionclip: Region-based language-image pretraining. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 16793–16803, 2022. 2

A Appendix

A.1 Regularization pseudocode def regularized_loss (Ims , Txts , lambda_patch , lambda_token ): # image _en cod er − ResNet −50 , t e x t _ e n c o d e r − CXR−BERT− s p e c i a l i z e d # Ims [ n , h , w, c ] , T x t s [ n , l ] − m i n i b a t c h o f a l i g n e d images & t e x t s # W_i [ d _ i , d _e ] − l e a r n e d p r o j e c t i o n s o f image p a t c h e s t o embed # W_t [ d _ t , d _e ] − l e a r n e d p r o j e c t i o n s o f t e x t t o k e n s t o embed # P − number o f image p a t ch e s , T − number o f t e x t t o k e n s # Se tup ; compute p a t ch and t o k en r e p r e s e n t a t i o n s patch_f = image_encoder (Ims) #[ n , d _ i , P ] token_f = text_encoder (Txts) #[ n , d _ t , T ] # p r o j e c t t o j o i n t embedd ing spa ce [ d _e ] and n o rm a l i z e patch_e = l2_normalize (dot(patch_f , W_i), axis =1) #[ n , d_e , P ] token_e = l2_normalize (dot(token_f , W_t), axis =1) #[ n , d_e , T ] # CLIP Lo s s ; Compute g l o b a l embedd ings image_e = l2_normalize (mean(patch_e , dim =2), axis =1) #[ n , d _e ] text_e = token_e [:, :, 0] #[ n , d _e ] # Compute s c a l e d p a i r w i s e c o s i n e s i m i l a r i t y m a t r i x clip_logits = dot(image_e , text_e.T) ∗ exp(t) #[ n , n ] # E v a l u a t e s ymme tr i c CLIP l o s s f u n c t i o n labels = np.arange(n) loss_i = cross_entropy_loss (clip_logits , labels , axis =0) loss_t = cross_entropy_loss (clip_logits , labels , axis =1) clip_loss = (loss_i + loss_t )/2 # R e g u l a r i z a t i o n ; Compute pa tch −t o k en s i m i l a r i t y m a t r i x sim_matrix = batch_multiply (token_e , patch_e) #[ n , T , P ] # Compute p a t ch and t o k en p e n a l t i e s patch_entropies = entropy(softmax(sim_matrix , axis = 2)) #[ n , T ] patch_penalty = lambda_patch ∗ mean( patch_entropies ) token_entropies = entropy(softmax(sim_matrix , axis = 1)) #[ n , P ] token_penalty = lambda_token ∗ mean( token_entropies ) regularized_loss = clip_loss + patch_penalty + token_penalty return regularized_loss Figure 6: Pseudocode for our TIER regularization method. lambda_patch and lambda_token are hyperparameters that can be tuned depending on the desired level of patch/text-token sparsity.

A.2 Additional Heatmap Examples

13_image_0.png

13_image_1.png

Figure 7: A CXR negative for cardiomegaly, overlaid with heatmaps displaying zero-shot cardiomegaly score for the unregularized (center) and regularized (right) models. Gray corresponds to a neutral (close to

zero) zero-shot score, while red is a higher score and blue is a lower score.

13_image_2.png

13_image_3.png

Figure 8: A CXR with atelectasis, overlayed with heatmaps displaying zero-shot atelectasis soore for the unregularized (center) and regularized (right) models.

13_image_4.png

Figure 9: A lateral view of the previous atelectasis-positive CXR. in Fig. 8, with zero-shot scores overlayed

A.3 Hyperparameter Sweep

Here, we present the results of the hyperparameter sweep we used to select our lambda hyperparameters.

Models were trained for a single epoch on MIMIC-CXR, and evaluated on the validation set of Chexpert. We use lambda values which maximized zero-shot AUC on the Chexpert validation set to train our regularized TIER model.

| λp | λt | λt = 0.00 | λt = 0.05 | λt = 0.10 | λt = 0.15 | λt = 0.20 | λt = 0.25 | |-----------|-------------|-------------|-------------|-------------|-------------|-------------| | λp = 0.00 | 0.84708 | 0.84514 | 0.84624 | 0.84272 | 0.84870 | 0.83311 | | λp = 0.05 | 0.85137 | 0.84661 | 0.84353 | 0.84542 | 0.84459 | 0.84712 | | λp = 0.10 | 0.83971 | 0.85457 | 0.85059 | 0.84736 | 0.84464 | 0.83879 | | λp = 0.15 | 0.83774 | 0.85367 | 0.85107 | 0.84174 | 0.83895 | 0.84242 | | λp = 0.20 | 0.84399 | 0.85387 | 0.85469 | 0.84488 | 0.84747 | 0.83234 | | λp = 0.25 | 0.84901 | 0.83306 | 0.84802 | 0.83939 | 0.83307 | 0.85160 |

Table 3: Zero-shot AUCs on the validation set after training a model with the given hyperparameters for 1 epoch on MIMIC-CXR. (λp = 0.20, λt = 0.10) were chosen to train our regularized TIER model.

A.4 P-Values For Chexpert Evaluation

Label CheXzero Mean CheXzero Std TIER (Ours) Mean TIER Std T statistic P value
Average AUC 0.893494 0.00696 0.903336 0.007213 31.05041349 4.5448E-173
Cardiomegaly 0.883397 0.013806 0.917135 0.010944 60.5584468 0
Edema 0.894235 0.015879 0.924225 0.01217 47.40345097 0
Consolidation 0.913176 0.014555 0.897116 0.023315 -18.47763326 1.67297E-70
Atelectasis 0.842985 0.015325 0.865306 0.014135 33.85648614 5.547E-199
Pleural Effusion 0.933676 0.010558 0.912898 0.012155 -40.81063492 2.4219E-265
Label Unreg. (Ours) Mean Unreg. Std TIER (Ours) Mean TIER Std T statistic P value
Average AUC 0.897206 0.007486 0.903336 0.007213 18.64716405 1.13364E-71
Cardiomegaly 0.892394 0.012546 0.917135 0.010944 46.99391106 0
Edema 0.907286 0.013561 0.924225 0.01217 29.39763772 3.6269E-158
Consolidation 0.912127 0.024044 0.897116 0.023315 -14.17328934 1.60039E-43
Atelectasis 0.857408 0.014522 0.865306 0.014135 12.32428706 1.08563E-33
Pleural Effusion 0.916813 0.011588 0.912898 0.012155 -7.372034693 2.44822E-13
------------------ ----------------- ---------------- ---------------------- -------------- --------------- -------------
Average AUC 0.893494 0.00696 0.897206 0.007486 11.48385375 1.32063E-29
Cardio 0.883397 0.013806 0.892394 0.012546 15.25117349 9.17475E-50
Edema 0.894235 0.015879 0.907286 0.013561 19.7641868 1.47602E-79
Consolidation 0.913176 0.014555 0.912127 0.024044 -1.180245623 0.238043032
Atelectasis 0.842985 0.015325 0.857408 0.014522 21.60293691 3.44293E-93
Pleural Effusion 0.933676 0.010558 0.916813 0.011588 -34.01616352 1.7778E-200

Here we present the p-values obtained from two-sample t-tests for differences in mean AUCs between TIER (Ours), CheXzero, and the unregularized baseline (Ours) for the chexpert evaluations. Table 4: Two sample t-test for the difference of means between TIER (Ours) and CheXzero for n = 1000 bootstraps of the chexpert evaluation. All results are significant at the p=0.0001 level.

Table 5: Two sample t-test for the difference of means between TIER (ours) and the unregularized baseline (ours) for n = 1000 bootstraps of chexpert evaluation. All results are significant at the p=0.0001 level. Table 6: Two sample t-test for the difference of means between unregularized (ours) and CheXzero for n = 1000 bootstraps of chexpert evaluation. All results but consolidation are significant at the p=0.0001 level.

A.5 Additional Chexpert Baselines

Here we display some additional baselines compared to our regularized TIER and unregularized baseline. In particular, we present the pretrained model we are using (Boecking et al., 2022) and CLIP (Radford et al., 2021).

Label TIER (Ours) Unregularized (Ours) BioViL CLIP
Average 0.903336 0.89721 0.63631 0.54056
Cardiomegaly 0.91714 0.89239 0.63300 0.56232
Edema 0.92423 0.90729 0.58706 0.5028
Consolidation 0.89712 0.91213 0.705898 0.6541
Atelectasis 0.86531 0.85741 0.58698 0.51719
Pleural Effusion 0.91290 0.91681 0.66864 0.46638

Table 7: The highest zero-shot AUC of the listed models is bolded. All models are performing zero-shot classification, meaning they have never explicitly been trained with any labels.

Class Label Caption Cardiomegaly is present.
Cardiomegaly The heart shadow is enlarged. The cardiac silhouette is enlarged.
Pleural Effusion Pleural Effusion is present. Blunting of the costophrenic angles represents pleural effusions. The pleural space is filled with fluid. Layering pleural effusions are present.
Edema Edema is present. Increased fluid in the alveolar wall indicates pulmonary edema.
Consolidation Consolidation is present. Dense white area of right lung indicative of consolidation.
Atelectasis Atelectasis is present. Basilar opacity and volume loss is likely due to atelectasis.
No Finding The lungs are clear. No abnormalities are present. The chest is normal. No clinically significant radiographic abnormalities. No radiographically visible abnormalities in the chest.

A.6 Chexpert Queries

Table 8: Query captions used for zero-shot classification. No Finding captions are used as the negative queries for classification on CheXpert, while the rest are used as positive queries for their respective labels.

A.7 Chexpert Radiologist Benchmarking

Here, we compare binary predictions from these zero-shot models against the predictions of 3 radiologists using MCC and F1 metrics. We selected optimal thresholds for MCC and F1 using their performance on the CheXpert validation set.

16_image_0.png

chextero Figure 10: Benchmarking thresholded scores against radiologists. MCC (Matthews correlation coefficient) scores and F1 scores for the three zero-shot models and three radiologists are shown. These zero-shot models are competitive with radiologists for most labels. While TIER often outperform both the unregularized baseline at CheXzero on both metrics, it is equivalent or worse in several cases.

A.8 P-Values For Padchest Evaluation

Here we present the p-values obtained from two-sample t-tests for differences in mean AUCs for n = 1000 bootstraps between the various models for the Padchest evaluations.

Comparing CheXzero to TIER (Ours) on Padchest head-to-head:

Label CheXzero Mean CheXzero Std TIER (Ours) Mean TIER Std T statistic P value
Average AUC 0.726306 0.002558 0.75542 0.002347 265.2017626 0
Number won 21 33
endotracheal tube 0.98295 0.002315 0.979606 0.005919 -16.63830145 2.56865E-58
pleural effusion 0.950519 0.002734 0.942045 0.002976 -66.3097905 0
pulmonary edema 0.95646 0.008905 0.941289 0.008719 -38.49466433 5.1755E-243
heart insufficiency 0.917819 0.004761 0.926097 0.005453 36.16180694 1.015E-220
pulmonary fibrosis 0.921793 0.008352 0.951654 0.0054 94.94482215 0
cardiomegaly 0.890467 0.002511 0.883692 0.002718 -57.89827507 0
vascular redistribution 0.750592 0.018104 0.877236 0.013662 176.5761336 0
consolidation 0.865175 0.009475 0.878342 0.009453 31.10977089 1.3088E-173
hilar congestion 0.825707 0.007219 0.855435 0.008915 81.95062909 0
pulmonary mass 0.842056 0.01203 0.844107 0.012986 3.663919249 0.000254856
cavitation 0.853367 0.01575 0.857639 0.015601 6.093824451 1.31926E-09
alveolar pattern 0.763811 0.006956 0.87631 0.005049 413.89493 0
calcified pleural thickening 0.850707 0.019816 0.859651 0.020362 9.95447536 8.11361E-23
lung metastasis 0.827675 0.024452 0.877375 0.015578 54.20861509 0
emphysema 0.830578 0.009475 0.717841 0.013142 -220.0452111 0
interstitial pattern 0.816432 0.005126 0.835144 0.005382 79.61342323 0
costophrenic angle blunting 0.69029 0.006777 0.769921 0.006022 277.7581639 0
tuberculosis 0.7978 0.027473 0.838961 0.020022 38.28895091 4.8878E-241
atelectasis 0.809232 0.009106 0.781707 0.008707 -69.08700011 0
reticular interstitial pattern 0.822429 0.021108 0.844479 0.022699 22.49540647 4.4633E-100
pneumonia 0.773941 0.005566 0.813796 0.004699 173.0195883 0
lobar atelectasis 0.775147 0.017958 0.808411 0.014991 44.96696255 1.1767E-305
normal 0.753171 0.002546 0.776328 0.003632 165.0977062 0
pleural thickening 0.752537 0.016871 0.784428 0.014559 45.25503582 0
reticulonodular interstitial pattern 0.841414 0.027064 0.862346 0.024737 18.05301834 1.31327E-67
infiltrates 0.747836 0.006206 0.742854 0.006681 -17.27714913 1.9175E-62
hypoexpansion 0.794564 0.014337 0.853423 0.011148 102.4871616 0
hypoexpansion basal 0.8018 0.015611 0.889652 0.013677 133.8542388 0
humeral fracture 0.749084 0.023222 0.742305 0.026582 -6.07337803 1.49532E-09
pneumothorax 0.777442 0.018202 0.730643 0.026112 -46.49431338 0
multiple nodules 0.716911 0.028257 0.790815 0.021245 66.10682237 0
hyperinflated lung 0.713202 0.018784 0.700879 0.018276 -14.8691189 1.64612E-47
bronchiectasis 0.690117 0.01037 0.734643 0.009854 98.42837415 0
adenopathy 0.703924 0.02035 0.678726 0.016664 -30.29508782 3.2002E-166
mediastinal enlargement 0.759299 0.022403 0.72538 0.026617 -30.83087996 4.5077E-171
laminar atelectasis 0.679276 0.006312 0.67343 0.006914 -19.74676158 1.97007E-79
vertebral compression 0.646448 0.025325 0.723955 0.018277 78.47811851 0
rib fracture 0.691037 0.020514 0.689835 0.022485 -1.248835772 0.211871446
tuberculosis sequelae 0.584302 0.019415 0.796895 0.013529 284.0954171 0
hilar enlargement 0.678564 0.012333 0.721779 0.011469 81.14282586 0
tracheal shift 0.634359 0.019985 0.615827 0.019305 -21.0906627 2.58516E-89
mediastinal mass 0.647695 0.034419 0.709825 0.031109 42.3483101 3.0928E-280
central vascular redistribution 0.354491 0.034431 0.728932 0.031302 254.4623065 0
vertebral fracture 0.499654 0.02921 0.791375 0.015662 278.3320678 0
superior mediastinal enlargement 0.596948 0.024252 0.551017 0.025206 -41.52441949 2.9881E-272
vascular hilar enlargement 0.623934 0.007406 0.625607 0.007007 5.189078029 2.32751E-07
nodule 0.546737 0.010556 0.446317 0.010124 -217.1150565 0
air trapping 0.580882 0.006245 0.580408 0.005897 -1.745118047 0.081118057
bullas 0.486494 0.020169 0.744606 0.018356 299.2954863 0
ground glass pattern 0.602802 0.027608 0.671321 0.020656 62.84105487 0
calcified adenopathy 0.583562 0.02315 0.673757 0.019228 94.77745188 0
minor fissure thickening 0.77315 0.018481 0.600411 0.025956 -171.4357871 0
unchanged 0.395541 0.004386 0.618171 0.004502 1120.102208 0
clavicle fracture 0.607514 0.036746 0.596974 0.037946 -6.309943613 3.42827E-10
pseudonodule 0.557981 0.009991 0.476977 0.011371 -169.2291715 0
end on vessel 0.560626 0.035794 0.397635 0.041243 -94.38338959 0
COPD signs 0.751217 0.003745 0.650859 0.004075 -573.421393 0

Table 9: Two sample t-test for the difference of means for our padchest evaluation. In this table, we compare the previous SOTA, CheXzero, with our regularized model, TIER. Each model had AUC evaluated on n = 1000 bootstraps of the radiologists-labeled subset of padchest. All but the air trapping, rib fracture, and pulmonary mass results are significant at the p=0.0001 level.

Label Unreg. (Ours) Mean Unreg. Std TIER (Ours) Mean TIER Std T statistic P value
Average AUC 0.742534 0.002567 0.75542 0.002347 117.1556311 0
Number won 23 30
endotracheal tube 0.956634 0.008621 0.979606 0.005919 69.46676879 0
pleural effusion 0.930292 0.003377 0.942045 0.002976 82.56984314 0
pulmonary edema 0.945415 0.008526 0.941289 0.008719 -10.69926209 5.11898E-26
heart insufficiency 0.927177 0.005282 0.926097 0.005453 -4.498643855 7.23167E-06
pulmonary fibrosis 0.944147 0.006584 0.951654 0.0054 27.87855697 9.3138E-145
cardiomegaly 0.883339 0.002701 0.883692 0.002718 2.913187348 0.00361733
vascular redistribution 0.872019 0.013502 0.877236 0.013662 8.588841321 1.73859E-17
consolidation 0.849872 0.011331 0.878342 0.009453 61.01092612 0
hilar congestion 0.850243 0.008691 0.855435 0.008915 13.18723777 3.87838E-38
pulmonary mass 0.872299 0.012346 0.844107 0.012986 -49.75455492 0
cavitation 0.794295 0.017229 0.857639 0.015601 86.18194266 0
alveolar pattern 0.816974 0.006339 0.87631 0.005049 231.535272 0
calcified pleural thickening 0.84287 0.019733 0.859651 0.020362 18.71497205 3.84423E-72
lung metastasis 0.860837 0.017219 0.877375 0.015578 22.52272401 2.7303E-100
emphysema 0.718377 0.013149 0.717841 0.013142 -0.911743481 0.362013758
interstitial pattern 0.840368 0.005014 0.835144 0.005382 -22.45845998 8.6714E-100
costophrenic angle blunting 0.808131 0.00461 0.769921 0.006022 -159.323741 0
tuberculosis 0.843741 0.024721 0.838961 0.020022 -4.751556022 2.1624E-06
atelectasis 0.791507 0.0086 0.781707 0.008707 -25.32275656 6.2416E-123
reticular interstitial pattern 0.867637 0.019332 0.844479 0.022699 -24.56163626 1.2365E-116
pneumonia 0.796614 0.005068 0.813796 0.004699 78.6172414 0
lobar atelectasis 0.815725 0.013775 0.808411 0.014991 -11.36063996 4.99873E-29
normal 0.790588 0.003023 0.776328 0.003632 -95.42795587 0
pleural thickening 0.754608 0.016022 0.784428 0.014559 43.55866303 5.5682E-292
reticulonodular interstitial pattern 0.838374 0.026992 0.862346 0.024737 20.70489002 1.95964E-86
infiltrates 0.735399 0.006646 0.742854 0.006681 25.01662683 2.1878E-120
hypoexpansion 0.871452 0.010845 0.853423 0.011148 -36.65734018 1.9617E-225
hypoexpansion basal 0.874477 0.014044 0.889652 0.013677 24.47917337 5.8658E-116
humeral fracture 0.672935 0.028067 0.742305 0.026582 56.74716788 0
pneumothorax 0.728547 0.021418 0.730643 0.026112 1.962595625 0.049831759
multiple nodules 0.852951 0.020427 0.790815 0.021245 -66.66997483 0
hyperinflated lung 0.667704 0.017846 0.700879 0.018276 41.06987413 7.5179E-268
bronchiectasis 0.743998 0.009203 0.734643 0.009854 -21.94072646 8.91339E-96
adenopathy 0.73105 0.017982 0.678726 0.016664 -67.49145771 0
mediastinal enlargement 0.666796 0.028092 0.72538 0.026617 47.87154657 0
laminar atelectasis 0.687839 0.006426 0.67343 0.006914 -48.27281362 0
vertebral compression 0.734413 0.019904 0.723955 0.018277 -12.2383364 2.91558E-33
rib fracture 0.668069 0.023081 0.689835 0.022485 21.36070437 2.38093E-91
tuberculosis sequelae 0.773832 0.014015 0.796895 0.013529 37.44003048 6.6415E-233
hilar enlargement 0.714687 0.011757 0.721779 0.011469 13.65450364 1.19318E-40
tracheal shift 0.500734 0.02327 0.615827 0.019305 120.3743682 0
mediastinal mass 0.409473 0.031929 0.709825 0.031109 213.0621769 0
central vascular redistribution 0.567387 0.046306 0.728932 0.031302 91.39738689 0
vertebral fracture 0.86009 0.013028 0.791375 0.015662 -106.6628934 0
superior mediastinal enlargement 0.637878 0.021595 0.551017 0.025206 -82.75530027 0
vascular hilar enlargement 0.60417 0.007237 0.625607 0.007007 67.29618289 0
nodule 0.507929 0.011538 0.446317 0.010124 -126.9283034 0
air trapping 0.631534 0.005968 0.580408 0.005897 -192.699814 0
bullas 0.584846 0.023316 0.744606 0.018356 170.2487729 0
ground glass pattern 0.661248 0.021925 0.671321 0.020656 10.57463045 1.81468E-25
calcified adenopathy 0.624151 0.023153 0.673757 0.019228 52.12228786 0
minor fissure thickening 0.558331 0.022571 0.600411 0.025956 38.68594962 7.5149E-245
unchanged 0.633874 0.004591 0.618171 0.004502 -77.22708382 0
clavicle fracture 0.596031 0.041522 0.596974 0.037946 0.530145578 0.596069909
pseudonodule 0.472281 0.010954 0.476977 0.011371 9.405369698 1.37144E-20
end on vessel 0.485072 0.037602 0.397635 0.041243 -49.54199726 0
COPD signs 0.652912 0.00394 0.650859 0.004075 -11.45351594 1.83491E-29

Comparing Unregularized baseline (Ours) to TIER (Ours) on Padchest head-to-head: Table 10: Two sample t-test for the difference of means for our padchest evaluation. In this table, we compare the previous SOTA, CheXzero, with our regularized model, TIER. Each model had AUC evaluated on n = 1000 bootstraps of the radiologists-labeled subset of padchest. All but the cardiomegaly, emphysema, pneumothorax, and clavicle fracture findings are significant at the p=0.0001 level.

A.9 Code/Model Availability

Code is available at https://github.com/apalepu13/TIER_Regularized_CLIP. For model checkpoints for the regularized/unregularized/fully supervised models, contact the authors. The BioViL model is available at https://huggingface.co/microsoft/BiomedVLP-CXR-BERT-specialized.