Explanation and Judgement of IR Ranking using LLM
| dc.contributor.author | Mondal, Santanu | |
| dc.date.accessioned | 2025-07-22T09:24:29Z | |
| dc.date.available | 2025-07-22T09:24:29Z | |
| dc.date.issued | 2024-06 | |
| dc.description | Dissertation under the supervision of Dr. Debapriyo Majumdar | en_US |
| dc.description.abstract | Pretrained transformer models such as BERT and T5 have significantly advanced the performance of information retrieval (IR) systems when fine-tuned with large-scale labeled datasets. However, their effectiveness diminishes notably in low-resource scenarios where annotated query-passage pairs are limited. This thesis explores an alternative supervision strategy by leveraging natural language explanations to enhance training signals during fine-tuning. We propose a novel methodology that augments traditional relevance labels with textual explanations generated by a large language model (LLM) using few-shot prompting. To achieve this, we generate explanations for 30,000 query-passage-label triples from the MS MARCO dataset using the open-source model google/gemma-2b, allowing for cost-free and scalable inference. These augmented samples are then used to fine-tune a T5-base sequence-to-sequence model, with the objective of producing both the relevance label and an accompanying explanation. During inference, the model predicts the label token, and the probability of that token is used as a soft relevance score, enabling efficient ranking. Empirical results demonstrate that our explanation-augmented retriever outperforms strong baselines, including BM25, a BERT reranker, and a T5 model trained with labels only. We further analyze the effectiveness of explanation order, training data size, and the quality of generated rationales. Our findings suggest that natural language explanations offer a powerful form of supervision, particularly valuable in data-scarce IR settings, and present a compelling direction for improving neural retrievers with minimal annotation overhead. | en_US |
| dc.identifier.citation | 46p. | en_US |
| dc.identifier.uri | http://hdl.handle.net/10263/7591 | |
| dc.language.iso | en | en_US |
| dc.publisher | Indian Statistical Institute, Kolkata | en_US |
| dc.relation.ispartofseries | MTech(CS) Dissertation;23-19 | |
| dc.subject | Information Retrieval | en_US |
| dc.subject | Natural Language Explanations | en_US |
| dc.subject | Large Language Models | en_US |
| dc.subject | T5 | en_US |
| dc.subject | MS MARCO | en_US |
| dc.subject | Sequence-to-Sequence Learning | en_US |
| dc.subject | Fine-Tuning | en_US |
| dc.title | Explanation and Judgement of IR Ranking using LLM | en_US |
| dc.type | Other | en_US |
Files
Original bundle
1 - 2 of 2
No Thumbnail Available
- Name:
- Santanu_dissertation_plag_check_report.pdf
- Size:
- 895.87 KB
- Format:
- Adobe Portable Document Format
- Description:
- Plagiarism_report
No Thumbnail Available
- Name:
- santanu_dissertation_signed.pdf
- Size:
- 695.97 KB
- Format:
- Adobe Portable Document Format
- Description:
- Dissertations - M Tech (CS)
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 1.71 KB
- Format:
- Item-specific license agreed upon to submission
- Description:
