diff --git a/paper/hubble_paperclip.tex b/paper/hubble_paperclip.tex index 52b05fa..5848e47 100644 --- a/paper/hubble_paperclip.tex +++ b/paper/hubble_paperclip.tex @@ -203,7 +203,7 @@ \section{Introduction} \label{fig:overview} \end{figure*} -\section{\changes{Related work}} +\section{\changes{Related Work}} % The CLIP family of foundation models, which in their original form embed images and associated captions into a common representation space via contrastive learning, have shown strong performance and generalization capabilities on a variety of downstream tasks including zero-shot classification and image retrieval~\citep{radford2021learning}. %