Explanation and Judgement of IR Ranking using LLM
No Thumbnail Available
Date
2024-06
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Indian Statistical Institute, Kolkata
Abstract
Pretrained transformer models such as BERT and T5 have significantly advanced the
performance of information retrieval (IR) systems when fine-tuned with large-scale
labeled datasets. However, their effectiveness diminishes notably in low-resource
scenarios where annotated query-passage pairs are limited. This thesis explores
an alternative supervision strategy by leveraging natural language explanations to
enhance training signals during fine-tuning. We propose a novel methodology that
augments traditional relevance labels with textual explanations generated by a large
language model (LLM) using few-shot prompting.
To achieve this, we generate explanations for 30,000 query-passage-label triples from
the MS MARCO dataset using the open-source model google/gemma-2b, allowing
for cost-free and scalable inference. These augmented samples are then used
to fine-tune a T5-base sequence-to-sequence model, with the objective of producing
both the relevance label and an accompanying explanation. During inference, the
model predicts the label token, and the probability of that token is used as a soft
relevance score, enabling efficient ranking.
Empirical results demonstrate that our explanation-augmented retriever outperforms
strong baselines, including BM25, a BERT reranker, and a T5 model trained
with labels only. We further analyze the effectiveness of explanation order, training
data size, and the quality of generated rationales. Our findings suggest that natural
language explanations offer a powerful form of supervision, particularly valuable
in data-scarce IR settings, and present a compelling direction for improving neural
retrievers with minimal annotation overhead.
Description
Dissertation under the supervision of Dr. Debapriyo Majumdar
Keywords
Information Retrieval, Natural Language Explanations, Large Language Models, T5, MS MARCO, Sequence-to-Sequence Learning, Fine-Tuning
Citation
46p.
