Skip to content

Commit 45b067a

Browse files
committed
Added papers to releated work
1 parent f556964 commit 45b067a

3 files changed

Lines changed: 46 additions & 12 deletions

File tree

report/main.bib

Lines changed: 23 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -35,3 +35,26 @@ @article{nogueira2019doc2query
3535
pages={2},
3636
year={2019}
3737
}
38+
39+
@article{pradeep2021expando,
40+
title={The expando-mono-duo design pattern for text ranking with pretrained sequence-to-sequence models},
41+
author={Pradeep, Ronak and Nogueira, Rodrigo and Lin, Jimmy},
42+
journal={arXiv preprint arXiv:2101.05667},
43+
year={2021}
44+
}
45+
46+
%mono
47+
@article{nogueira2020document,
48+
title={Document ranking with a pretrained sequence-to-sequence model},
49+
author={Nogueira, Rodrigo and Jiang, Zhiying and Lin, Jimmy},
50+
journal={arXiv preprint arXiv:2003.06713},
51+
year={2020}
52+
}
53+
54+
%duo
55+
@article{nogueira2019multi,
56+
title={Multi-stage document ranking with BERT},
57+
author={Nogueira, Rodrigo and Yang, Wei and Cho, Kyunghyun and Lin, Jimmy},
58+
journal={arXiv preprint arXiv:1910.14424},
59+
year={2019}
60+
}

report/main.pdf

21.1 KB
Binary file not shown.

report/main.tex

Lines changed: 23 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -73,18 +73,28 @@ \section{Problem Statement}\label{sec:problem}
7373
\section{Related Work}\label{sec:related}
7474
In this section, we delve into pertinent research encompassing the realms of conversational search engines and the broader area of information retrieval. While certain highlighted studies do not directly cater to conversational search engines or explicit information retrieval, their techniques remain invaluable in various stages of the conversational retrieval process.
7575

76-
\subsection*{doc2query}
76+
\subsection*{\texttt{RM3} Pseudo-Relevance Feedback Query Expansion}
7777

78-
\subsection*{Text-to-Text Transfer Transformer}
78+
\subsection*{Text-to-Text Transfer Transformer}\label{sec:t5}
7979
The vast domain of natural language processing (NLP) revolves around the understanding of natural language, whether presented in text or speech form. NLP aspires to equip computers with the capability to grasp the depth of human language and harness this understanding to execute a range of tasks, such as text summarization, machine translation, and question answering. Given the diverse nature of these tasks in terms of their input, output, and underlying challenges, developing a unified model proficient across the entire spectrum poses a significant challenge.
8080

81-
Enter the Text-to-Text Transfer Transformer (T5) \cite{raffel2020exploring}. This work by Raffel et al. introduces transfer learning in NLP, aiming to craft a versatile model that can be used for any NLP problem. In essence, T5 models first learn the basics of language. Then, they're sharpened for particular tasks using targeted data. It's common to find models that have been trained in this manner for any specific NLP problem.
81+
Enter the Text-to-Text Transfer Transformer (\texttt{T5}) \cite{raffel2020exploring}. This work by Raffel et al. introduces transfer learning in NLP, aiming to craft a versatile model that can be used for any NLP problem. In essence, T5 models first learn the basics of language. Then, they're sharpened for particular tasks using targeted data. It's common to find models that have been trained in this manner for any specific NLP problem.
8282

83-
\subsection*{doc2query-T5}
83+
\subsection*{\texttt{doc2query}}
84+
Traditional retrieval techniques, such as \texttt{BM25}, rely primarily on term occurrences in both queries and documents. However, they often overlook the semantics of the content. As a result, documents that may be semantically relevant to a query might be scored as non-relevant due to differences in syntax or terminology. Dense retrieval methods, which emphasize semantic similarities between texts, can address this problem but are computationally taxing during retrieval.
8485

85-
\subsection*{monoT5 \& duoT5}
86+
A notable solution to this is the \texttt{doc\-2query} method proposed by Nogueira et al. \cite{nogueira2019document}. It employs a text-to-text transformer to convert documents into queries. By generating and appending a few of these transformed queries to the original document, classical retrieval methods show significantly improved performance. This is because these additional queries often capture semantic nuances similar to those in the actual query \cite{nogueira2019document,nogueira2019doc2query,pradeep2021expando}. Importantly, \texttt{doc\-2query} shifts the computational load to the indexing phase, ensuring minimal performance lag during retrieval. By leveraging the \texttt{T5} model, the authors further enhanced the query generation quality, leading to the variation known as \texttt{doc\-TTTTTquery}, \texttt{doc\--T5query}, or \texttt{doc\-2query\--T5} \cite{nogueira2019doc2query}.
8687

87-
\subsection*{T5 Query Rewriting}
88+
\subsection*{\texttt{monoT5} \& \texttt{duoT5} Rerankers}
89+
\texttt{monoT5} and \texttt{duoT5} are neural re-rankers, also developed by Nogueira et al., which attempt to inject semantic understanding into the retrieval process \cite{nogueira2020document,nogueira2019multi}. Using the \texttt{T5} model, they re-rank a list of documents based on their semantic relevance to a given query. Specifically, \texttt{monoT5} processes a query and a single document, outputting a relevance score. In contrast, \texttt{duoT5} considers a query and two documents, determining which document is more relevant. Although \texttt{duoT5} offers a more nuanced ranking, its pairwise comparison method makes it computationally heavier. Hence, a staged re-ranking approach is proposed: first using \texttt{monoT5} for the top $k$ documents and subsequently applying \texttt{duoT5} to a smaller subset, the top $l$, where $l \ll k$ \cite{nogueira2019multi,pradeep2021expando}.
90+
91+
\subsection*{\texttt{SPARTA} Reranker}
92+
93+
94+
\subsection*{Expando-Mono-Duo Design Pattern}
95+
The same research team introduced a strategic pattern for integrating the above tools into retrieval pipelines, termed the Expando-Mono-Duo design pattern \cite{pradeep2021expando}. Here's how it works: During indexing, \texttt{doc2query-T5} is employed to enhance document representation and better the initial retrieval results from methods like \texttt{BM25}. The retrieved results are then re-ranked with \texttt{monoT5}. A selected top tier from this list undergoes another re-ranking using \texttt{duoT5}. Trials show that this composite approach leads to marked improvements in result quality across multiple evaluation metrics \cite{pradeep2021expando}.
96+
97+
\subsection*{\texttt{T5} Conversational Query Rewriting}
8898

8999

90100

@@ -112,22 +122,22 @@ \subsection{T5 Query Rewriting}
112122

113123
We have found through experimentation that this mere concatenation is insufficient to produce satisfiable results: The concatenated string is too long, and too much focus during the later retrieval is being put on $r_{n-1}$, which make responses sudden topic changes impossible.
114124

115-
Driven by these insights, we have turned our attention to other query rewriting techniques. \texttt{Pyterrier} provides the \texttt{SequentialDependence} query rewriting method\footnote{URL: \url{https://pyterrier.readthedocs.io/en/latest/rewrite.html\#sequentialdependence}}. We have found, however, that this rewriter also does not produce the desired results.
125+
Driven by these insights, we have turned our attention to other query rewriting techniques. \texttt{Pyterrier} provides the \texttt{Sequential\-Dependence} query rewriting method\footnote{URL: \url{https://pyterrier.readthedocs.io/en/latest/rewrite.html\#sequentialdependence}}. We have found, however, that this rewriter also does not produce the desired results.
116126

117127
Subsequent exploration led us to the \texttt{T5} neural query rewriter trained for conversational question rewriting\footnote{URL: \url{https://huggingface.co/castorini/t5-base-canard}}. With this method, $q'_n$ closely mirrored $q_n$, subtly infusing it with the conversation's context, particularly when no drastic topic alterations were identified. A valuable by-product was the concise nature of the rewritten query, a departure from the growing length observed previously.
118128

119129
\subsection{BM25 Retrieval}
120130
We settled on the \texttt{BM25} retrieval method, a commonly used formula in the realm of information retrieval, for its simplicity and its deployment in the reference system, allowing for direct comparisons.
121131

122132
\subsection{Re-ranking}
123-
The re-ranking stage of our baseline system consists of two stages: First, the top 1000 documents retrieved by the \texttt{BM25} retrieval method are re-ranked using the \texttt{monoT5} reranker. Afterwards, the top 50 documents of the previous re-ranking stage are rearranged using the \texttt{duoT5} reranker. e precise count of documents subject to reranking at each stage is a hyperparameter of our system, allowing to balance computational cost and result quality. These rerankers were implemented in the \texttt{pyterrier\_t5} library.\footnote{URL: \url{https://github.com/terrierteam/pyterrier_t5}}
133+
The re-ranking stage of our baseline system consists of two stages: First, the top 1000 documents retrieved by the \texttt{BM25} retrieval method are re-ranked using the \texttt{monoT5} reranker. Afterwards, the top 50 documents of the previous re-ranking stage are rearranged using the \texttt{duoT5} reranker. The precise count of documents subject to reranking at each stage is a hyperparameter of our system, allowing to balance computational cost and result quality. These rerankers were implemented in the \texttt{pyterrier\_t5} library.\footnote{URL: \url{https://github.com/terrierteam/pyterrier_t5}}
124134

125135

126136
\section{Advanced Method}\label{sec:advanced}
127137
Explain what you are taking as your advanced method(s), as well as why this is a promising attempt to outperform the baseline method, and why you are making specific implementation choices.
128138

129139
\section{Results}\label{sec:results}
130-
The individual methods were evaluated on the MS MARCO document collection and the following provided files: \texttt{queries\_train.csv}, comprising of a list of queries grouped together into several conversation sessions, and \texttt{qrels\_train.txt} that contains the relevance assessments for the training queries. Our evaluation focused on a suite of metrics:
140+
The individual methods were evaluated on the MS MARCO document collection and the following provided files: \texttt{queries\-\_\-train\-.csv}, comprising of a list of queries grouped together into several conversation sessions, and \texttt{qrels\-\_\-train\-.txt} that contains the relevance assessments for the training queries. Our evaluation focused on a suite of metrics:
131141
\begin{itemize}
132142
\item Recall at 1000 (R@1000)
133143
\item Mean Average Precision (MAP)
@@ -147,10 +157,11 @@ \section{Results}\label{sec:results}
147157
\begin{center}
148158
\caption{Performance of the different methods on the MS MARCO document collection.}
149159
\begin{tabular}{l|rrrr}
150-
& R@1000 & MAP & MRR & NDCG\_Cut@3 \\
160+
& MAP & MRR & R@1000 & NDCG\_Cut@3 \\
151161
\hline
152-
Reference System & ??? & 0.07 & ??? & 0.09 \\
153-
Baseline Method & $0.1746$ & $0.3230$ & $0.5705$ & $0.2461$ \\
162+
Reference System & 0.07 & & & 0.09 \\
163+
Baseline & $0.1746$ & $0.3230$ & $0.5705$ & $0.2461$ \\
164+
Baseline + \texttt{RM3} & $\mathbf{0.2124}$ & $\mathbf{0.3600}$ & $\mathbf{0.5710}$ & $\mathbf{0.2673}$
154165
\end{tabular}
155166
\label{table:1}
156167
\end{center}

0 commit comments

Comments
 (0)