Skip to content

Commit

Permalink
Browse files Browse the repository at this point in the history
  • Loading branch information
smsharma committed Jan 14, 2024
2 parents 6ec1de7 + 386170a commit 00a5d4d
Showing 1 changed file with 13 additions and 1 deletion.
14 changes: 13 additions & 1 deletion paper/main.tex
Original file line number Diff line number Diff line change
Expand Up @@ -241,7 +241,7 @@ \section{Methodology}

\paragraph*{Evaluation}

The model is evaluated by tracking the loss in Eq.~\ref{eq:softmax_loss} as well as the top-$k\%$ retrieval accuracy on the held out validation set over the course of training. The retrieval accuracy is defined as the fraction of associated captions which fall within the top $k\%$ of captions by cosine similarity of the (normalized) embeddings $(x_i \cdot y_j)$, averaged over the images in the validation set.
The model is evaluated by tracking the loss in Eq.~\ref{eq:softmax_loss} as well as the top-$k\%$ retrieval accuracy on the held out validation set over the course of training. The retrieval accuracy is defined as the fraction of associated captions which fall within the top $k\%$ of captions by cosine similarity of the (normalized) embeddings $x_i \cdot y_j$, averaged over the images in the validation set.

We also qualitatively evaluate the learned embeddings through image retrieval (i.e., retrieving the most relevant images from a set using natural language queries) and description retrieval (i.e., querying the astrophysical object classes and science use cases most relevant to a given observation) experiments.

Expand Down Expand Up @@ -276,6 +276,11 @@ \section{Results and Discussion}

The top 3 text associations are shown for each image query. The `ground truth' summarized abstract is shown in the right column. The base model is seen to return a mix of relevant and less-relevant associations. While it can often return the nature of objects imaged, we observe it to seldom return scientific phenomena (e.g., `dark matter' as successfully done by the fine-tuned model in the second row). The third row (supernova 1987A) highlights interesting behavior -- the base model erroneously attributes the object at the center of the image to a gravitational lens or protoplanetary disk, while the fine-tuned model correctly identifies it as a supernova remnant (which play a crucial role for interstellar chemistry -- another returned snippet).

\paragraph*{Cosine similarity distribution}

The top row of Figure~\ref{fig:sim_valtrain} shows the distribution of cosine similarities between corresponding image and text embeddings, $x_i$ and $y_i$, for the LLM-summarized abstracts using the fine-tuned CLIP model (red line), with the ordering of text embeddings shuffled (dashed red line), and for the base CLIP model (blue line), shown for the set of validation pairs (top left) and same number of training set pairs (top right). The distribution for the base model is seen to be sharply peaked at a specific value, showing little diversity and being



\begin{figure*}[!h]
\includegraphics[width=0.95\textwidth]{plots/tti_base.pdf}
Expand All @@ -295,6 +300,13 @@ \section{Results and Discussion}
\label{fig:itt}
\end{figure*}

\begin{figure*}[!h]
\includegraphics[width=0.95\textwidth]{plots/sim_valtrain.pdf}
\centering\includegraphics[width=0.45\textwidth]{plots/sim_summ1.pdf}
\caption{Distribution of cosine similarities between corresponding image and text embeddings, $x_i$ and $y_i$. (Top row) For the LLM-summarized abstracts using the fine-tuned CLIP model (red line), with the ordering of text embeddings shuffled (dashed red line), and for the base CLIP model (blue line), shown for (top left) the set of validation pairs and (top right) same number of training set pairs. (Bottom) For the single-concept summaries, when evaluated using the CLIP model fine-tuned on LLM-summarized abstracts (red) and fine-tuned on the single-concept summaries themselves.}
\label{fig:sim_valtrain}
\end{figure*}

\section{Outlook and Conclusions}
\label{sec:conclusion}

Expand Down

0 comments on commit 00a5d4d

Please sign in to comment.