HLTCOE at LiveRAG: GPT-Researcher using ColBERT retrieval
Abstract
The HLTCOE LiveRAG submission employed the GPT-researcher framework, ColBERT bi-encoder, FineWeb10-BT collection, PLAID-X, Qwen2.5-7B-Instruct, m2-bert-80M-8k-retrieval, and Falcon3-10B to achieve 5th place in the LiveRAG automatic evaluation for correctness.
The HLTCOE LiveRAG submission utilized the GPT-researcher framework for researching the context of the question, filtering the returned results, and generating the final answer. The retrieval system was a ColBERT bi-encoder architecture, which represents a passage with many dense tokens. Retrieval used a local, compressed index of the FineWeb10-BT collection created with PLAID-X, using a model fine-tuned for multilingual retrieval. Query generation from context was done with Qwen2.5-7B-Instruct, while filtering was accomplished with m2-bert-80M-8k-retrieval. Up to nine passages were used as context to generate an answer using Falcon3-10B. This system placed 5th in the LiveRAG automatic evaluation for correctness with a score of 1.07.
Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper