You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: report/main.bib
+35Lines changed: 35 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -87,3 +87,38 @@ @article{thakur2021beir
87
87
journal={arXiv preprint arXiv:2104.08663},
88
88
year={2021}
89
89
}
90
+
91
+
@article{vaidyanathan2015query,
92
+
title={Query expansion strategy based on pseudo relevance feedback and term weight scheme for monolingual retrieval},
93
+
author={Vaidyanathan, Rekha and Das, Sujoy and Srivastava, Namita},
94
+
journal={arXiv preprint arXiv:1502.05168},
95
+
year={2015}
96
+
}
97
+
98
+
@article{abdul2004umass,
99
+
title={UMass at TREC 2004: Novelty and HARD},
100
+
author={Abdul-Jaleel, Nasreen and Allan, James and Croft, W Bruce and Diaz, Fernando and Larkey, Leah and Li, Xiaoyan and Smucker, Mark D and Wade, Courtney},
101
+
journal={Computer Science Department Faculty Publication Series},
102
+
pages={189},
103
+
year={2004}
104
+
}
105
+
106
+
@inproceedings{lavrenko2017relevance,
107
+
title={Relevance-based language models},
108
+
author={Lavrenko, Victor and Croft, W Bruce},
109
+
booktitle={ACM SIGIR Forum},
110
+
volume={51},
111
+
number={2},
112
+
pages={260--267},
113
+
year={2017},
114
+
organization={ACM New York, NY, USA}
115
+
}
116
+
117
+
@article{chen2022pseudo,
118
+
title={Pseudo-Relevance Feedback Method Based on the Topic Relevance Model},
119
+
author={Chen, Zhaofeng and Guo, Naixuan and Sun, Jiu and Wang, Yuanyuan and Zhou, Feng and Xu, Sen and Wang, Rugang},
In this Section, we delve into pertinent research encompassing the realms of conversational search engines and the broader area of information retrieval. While certain highlighted studies do not directly cater to conversational search engines or explicit information retrieval, their techniques remain invaluable in various stages of the conversational retrieval process.
75
75
76
76
\subsection*{Pseudo-Relevance Feedback by Query Expansion}\label{sec:prf}
77
+
Oftentimes, queries entered by users are too short and include too little hint for the search engine to retrieve all relevant documents \cite{vaidyanathan2015query}. To counter this, the user-generated query must be reformulated to better reflect the user's informational needs. One approach to this is known as query expansion using relevance feedback: Assume we have a set of documents relevant to the user-generated query. One can then reformulate the query such that the retrieval system retrieves documents similar to the set of relevant documents, i.e. documents that are probably also relevant to the user's query. This expansion is accomplished by selecting keywords from the set of relevant documents and appending those to the query. Different strategies for this exist \cite{vaidyanathan2015query}.
78
+
79
+
However, in a real setting, we do not have a set of relevant documents. Instead, we can generate a set of documents that are hopefully relevant to the user's query, and then follow the same process. Since the relevance of the generated set of documents is unknown, this approach is called pseudo-relevance feedback. In order to retrieve this set, we can use the original user entered query to retrieve only the top $p$ documents. If $p$ is very small, it is likely that all those documents are relevant to the query. Expanding the query using those documents will then likely result in a better retrieval of all relevant documents \cite{vaidyanathan2015query}.
80
+
81
+
One particular technique for this is known as \texttt{RM3} \cite{abdul2004umass}. \texttt{RM3} uses relevance-based language models \cite{lavrenko2017relevance} to estimate the relevance of a word based on the word's probability in the language models of the (pseudo-)relevant documents. This relevance of the word to the query is then interpolated with the original query language model to avoid the query drifting too far away from the original query. The resulting language model is then used to generate a new query. Overall, the \texttt{RM3} approach can be parameterized in the number of documents to use for the pseudo-relevance feedback, the number of terms to add to the query, and the interpolation parameter \cite{abdul2004umass,lavrenko2017relevance,chen2022pseudo}.
82
+
77
83
78
84
\subsection*{Text-to-Text Transfer Transformer}\label{sec:t5}
79
85
The vast domain of natural language processing (NLP) revolves around the understanding of natural language, whether presented in text or speech form. NLP aspires to equip computers with the capability to grasp the depth of human language and harness this understanding to execute a range of tasks, such as text summarization, machine translation, and question answering. Given the diverse nature of these tasks in terms of their input, output, and underlying challenges, developing a unified model proficient across the entire spectrum poses a significant challenge.
0 commit comments