Explanation and Judgement of IR Ranking using LLM

dc.contributor.authorMondal, Santanu
dc.date.accessioned2025-07-22T09:24:29Z
dc.date.available2025-07-22T09:24:29Z
dc.date.issued2024-06
dc.descriptionDissertation under the supervision of Dr. Debapriyo Majumdaren_US
dc.description.abstractPretrained transformer models such as BERT and T5 have significantly advanced the performance of information retrieval (IR) systems when fine-tuned with large-scale labeled datasets. However, their effectiveness diminishes notably in low-resource scenarios where annotated query-passage pairs are limited. This thesis explores an alternative supervision strategy by leveraging natural language explanations to enhance training signals during fine-tuning. We propose a novel methodology that augments traditional relevance labels with textual explanations generated by a large language model (LLM) using few-shot prompting. To achieve this, we generate explanations for 30,000 query-passage-label triples from the MS MARCO dataset using the open-source model google/gemma-2b, allowing for cost-free and scalable inference. These augmented samples are then used to fine-tune a T5-base sequence-to-sequence model, with the objective of producing both the relevance label and an accompanying explanation. During inference, the model predicts the label token, and the probability of that token is used as a soft relevance score, enabling efficient ranking. Empirical results demonstrate that our explanation-augmented retriever outperforms strong baselines, including BM25, a BERT reranker, and a T5 model trained with labels only. We further analyze the effectiveness of explanation order, training data size, and the quality of generated rationales. Our findings suggest that natural language explanations offer a powerful form of supervision, particularly valuable in data-scarce IR settings, and present a compelling direction for improving neural retrievers with minimal annotation overhead.en_US
dc.identifier.citation46p.en_US
dc.identifier.urihttp://hdl.handle.net/10263/7591
dc.language.isoenen_US
dc.publisherIndian Statistical Institute, Kolkataen_US
dc.relation.ispartofseriesMTech(CS) Dissertation;23-19
dc.subjectInformation Retrievalen_US
dc.subjectNatural Language Explanationsen_US
dc.subjectLarge Language Modelsen_US
dc.subjectT5en_US
dc.subjectMS MARCOen_US
dc.subjectSequence-to-Sequence Learningen_US
dc.subjectFine-Tuningen_US
dc.titleExplanation and Judgement of IR Ranking using LLMen_US
dc.typeOtheren_US

Files

Original bundle

Now showing 1 - 2 of 2
No Thumbnail Available
Name:
Santanu_dissertation_plag_check_report.pdf
Size:
895.87 KB
Format:
Adobe Portable Document Format
Description:
Plagiarism_report
No Thumbnail Available
Name:
santanu_dissertation_signed.pdf
Size:
695.97 KB
Format:
Adobe Portable Document Format
Description:
Dissertations - M Tech (CS)

License bundle

Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Item-specific license agreed upon to submission
Description: