Papers
arxiv:2506.22356

HLTCOE at LiveRAG: GPT-Researcher using ColBERT retrieval

Published on Jun 27
Authors:
,
,
,
,

Abstract

The HLTCOE LiveRAG submission employed the GPT-researcher framework, ColBERT bi-encoder, FineWeb10-BT collection, PLAID-X, Qwen2.5-7B-Instruct, m2-bert-80M-8k-retrieval, and Falcon3-10B to achieve 5th place in the LiveRAG automatic evaluation for correctness.

AI-generated summary

The HLTCOE LiveRAG submission utilized the GPT-researcher framework for researching the context of the question, filtering the returned results, and generating the final answer. The retrieval system was a ColBERT bi-encoder architecture, which represents a passage with many dense tokens. Retrieval used a local, compressed index of the FineWeb10-BT collection created with PLAID-X, using a model fine-tuned for multilingual retrieval. Query generation from context was done with Qwen2.5-7B-Instruct, while filtering was accomplished with m2-bert-80M-8k-retrieval. Up to nine passages were used as context to generate an answer using Falcon3-10B. This system placed 5th in the LiveRAG automatic evaluation for correctness with a score of 1.07.

Community

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2506.22356 in a model README.md to link it from this page.

Datasets citing this paper 1

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2506.22356 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.