Proceedings - NeuCLIR 2024¶
IRLab-AMS at TREC’24 NeuCLIR Track¶
Andrew Yates, Jia-Huei Ju
- Participant: IRLab-Amsterdam
- Paper: https://trec.nist.gov/pubs/trec33/papers/IRLab-Amsterdam.neuclir.pdf
- Runs: mlir-IRLabAmsterdam-ANEL-titledesc | mlir-IRLabAmsterdam-ANEL-desc | mlir-IRLabAmsterdam-ANEL-title | rus_irlab-ams-std-translate-llama-70B-api | zho_irlab-ams-std-translate-llama-70B-api | fas_irlab-ams-std-translate-llama-70B-api | fas_irlab-ams-postcite-v | rus_irlab-ams-postcite-v | zho_irlab-ams-postcite-v | zho_irlab-ams-std-recomp-llama-8B | rus_irlab-ams-std-recomp-llama-8B | fas_irlab-ams-std-recomp-llama-8B | fas_irlab-ams-postcite | rus_irlab-ams-postcite | zho_irlab-ams-postcite | fas_irlab-ams-std-translate-llama-8B | rus_irlab-ams-std-translate-llama-8B | zho_irlab-ams-std-translate-llama-8B | zho_irlab-ams-std-mdcomp-330-translate-llama-8B | rus_irlab-ams-std-mdcomp-330-translate-llama-8B | fas_irlab-ams-std-mdcomp-330-translate-llama-8B | fas_irlab-ams-std-mdcomp-331-translate-llama-8B | rus_irlab-ams-std-mdcomp-331-translate-llama-8B | zho_irlab-ams-std-mdcomp-331-translate-llama-8B
Abstract
In this notebook paper, we describe our participation as IRLab-AMS in the NeuCLIR. Our submitted results for two tasks, multi-lingual information retrieval (MLIR) and cross-language report generation (ReportGen). For MLIR, we explore the learned sparse representations with multi-lingual retrieval settings. For ReportGen, we experiment with several pipelines for generating long-form reports, including standard retrieval-augmented generation (RAG) and post-hoc citation methods. Additionally, we add an extra retrieval augmentation module to handle the limitation of ad-hoc retriever. The module can serve as distinct purposes, including relevance ranking, novelty ranking, and summarization, or by combining them.
Bibtex
@inproceedings{IRLab-Amsterdam-trec2024-papers-proc-1,
title = {IRLab-AMS at TREC’24 NeuCLIR Track},
author = {Andrew Yates and Jia-Huei Ju},
booktitle = {Proceedings of the 33th Text {REtrieval} Conference (TREC 2024)},
year = {2024},
address = {Gaithersburg, Maryland},
series = {NIST SP 1329}
}
Overview of the TREC 2024 NeuCLIR Track¶
Dawn Lawrie, Sean MacAvaney, James Mayfield, Paul McNamee, Douglas W. Oard, Luca Soldaini, Eugene Yang
- Participant: hltcoe
- Paper: https://trec.nist.gov/pubs/trec33/papers/hltcoe.neuclir.pdf
- Runs: fas-hltcoe-MNED-plaid_distill_engeng_zs2engfas | fas-hltcoe-MNED-plaid_distill_engeng_zs2fasfas | fas-hltcoe-MNED-plaid_distill_engfas | fas-hltcoe-MNED-plaid_distill_engfas.mt5rerank.gpt4rerank | fas-hltcoe-MNED-plaid_distill_engfas_450p | fas-hltcoe-MNED-plaid_distill_engfas_termpool2 | fas-hltcoe-MNED-plaid_distill_engmlir | fas-hltcoe-MNED-plaid_distill_engfas.mt5rerank | fas-hltcoe-MNED-plaid_distill_mono_fasfas | fas-hltcoe-MNED-plaid_eqsynms_distill_engfas | fas-hltcoe-MNED-plaid_syn_distill_engfas | fas-hltcoe-MNTEH-kitchen_rankfuse | fas-hltcoe-MNTEH-kitchen_rankfuse,mt5rerank | fas-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.gpt4rerank | fas-hltcoe-MTED-plaid_distill_engeng | rus-hltcoe-MNED-plaid_distill_engeng_zs2engrus | rus-hltcoe-MNED-plaid_distill_engeng_zs2rusrus | rus-hltcoe-MNED-plaid_distill_engmlir | rus-hltcoe-MNED-plaid_distill_engrus | rus-hltcoe-MNED-plaid_distill_engrus.mt5rerank | rus-hltcoe-MNED-plaid_distill_engrus.mt5rerank.gpt4rerank | rus-hltcoe-MNED-plaid_distill_engrus_450p | rus-hltcoe-MNED-plaid_distill_engrus_termpool2 | rus-hltcoe-MNED-plaid_distill_mono_rusrus | rus-hltcoe-MNED-plaid_eqsynms_distill_engrus | rus-hltcoe-MNED-plaid_syn_distill_engrus | rus-hltcoe-MNTEH-kitchen_rankfuse | rus-hltcoe-MNTEH-kitchen_rankfuse,mt5rerank | rus-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.gpt4rerank | rus-hltcoe-MTED-plaid_distill_engeng | zho-hltcoe-MNED-plaid_distill_engeng_zs2engzho | zho-hltcoe-MNED-plaid_distill_engeng_zs2zhozho | zho-hltcoe-MNED-plaid_distill_engmlir | zho-hltcoe-MNED-plaid_distill_engzho | zho-hltcoe-MNED-plaid_distill_engzho.mt5rerank | zho-hltcoe-MNED-plaid_distill_engzho.mt5rerank.gpt4rerank | zho-hltcoe-MNED-plaid_distill_engzho_450p | zho-hltcoe-MNED-plaid_distill_engzho_termpool2 | zho-hltcoe-MNED-plaid_distill_mono_zhozho | zho-hltcoe-MNED-plaid_eqsynms_distill_engzho | zho-hltcoe-MNED-plaid_syn_distill_engzho | zho-hltcoe-MNTEH-kitchen_rankfuse | zho-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank | zho-hltcoe-MTED-plaid_distill_engeng | zho-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.gpt4rerank | tech-hltcoe-MNED-plaid_distill_engeng_zs2engzho | tech-hltcoe-MNED-plaid_distill_engeng_zs2zhozho | tech-hltcoe-MNED-plaid_distill_engzho | tech-hltcoe-MNED-plaid_distill_zhozho | tech-hltcoe-MNED-plaid_syn_distill_engzho | tech-hltcoe-MNTEH-kitchen_rankfuse | tech-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank | tech-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.gpt4rerank | tech-hltcoe-MTED-plaid_distill_engeng | tech-hltcoe-MNED-plaid_eqsynms_distill_engzho | mlir-hltcoe-MNED-plaid_distill_clir_scorefuse | mlir-hltcoe-MNED-plaid_distill_clir.mt5rerank.scorefuse | mlir-hltcoe-MNED-plaid_distill_clir.mt5rerank.scorefuse.gpt4rerank | mlir-hltcoe-MNED-plaid_distill_engeng_zs | mlir-hltcoe-MNED-plaid_distill_mlir_bycoll_scorefuse | mlir-hltcoe-MNED-plaid_distill_mlir_mixedentry | mlir-hltcoe-MNED-plaid_distill_mlir_mixedentry_termpool2 | mlir-hltcoe-MNED-plaid_distill_mlir_mixedpass | mlir-hltcoe-MNED-plaid_distill_mlir_rr | mlir-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.scorefuse | mlir-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.scorefuse.gpt4rerank | mlir-hltcoe-MTED-plaid_distill_engeng | zho-jhu-orion-aggregated-w-gpt4o | rus-jhu-orion-aggregated-w-gpt4o | fas-jhu-orion-aggregated-w-gpt4o | zho-jhu-orion-aggregated-w-claude | rus-jhu-orion-aggregated-w-claude | fas-jhu-orion-aggregated-w-claude | fas-hltcoe-eugene-gpt4o | rus-hltcoe-eugene-gpt4o | fas-hltcoe-eugene-gpt35turbo | rus-hltcoe-eugene-gpt35turbo | zho-hltcoe-eugene-gpt35turbo | zho-hltcoe-eugene-gpt4o-fixed
Abstract
The principal goal of the TREC Neural Cross-Language Information Retrieval (NeuCLIR) track is to study the effect of neural approaches on cross-language information access. The track has created test collections containing Chinese, Persian, and Russian news stories and Chinese academic abstracts. NeuCLIR includes four task types: Cross-Language Information Retrieval (CLIR) from news, Multilingual Information Retrieval (MLIR) from news, Report Generation from news, and CLIR from technical documents. A total of 274 runs were submitted by five participating teams (and as baselines by the track coordinators) for eight tasks across these four task types. Task descriptions and the available results are presented.
Bibtex
@inproceedings{hltcoe-trec2024-papers-proc-2,
title = {Overview of the TREC 2024 NeuCLIR Track},
author = {Dawn Lawrie and Sean MacAvaney and James Mayfield and Paul McNamee and Douglas W. Oard and Luca Soldaini and Eugene Yang},
booktitle = {Proceedings of the 33th Text {REtrieval} Conference (TREC 2024)},
year = {2024},
address = {Gaithersburg, Maryland},
series = {NIST SP 1329}
}
HLTCOE at TREC 2024 NeuCLIR Track¶
Eugene Yang, Dawn Lawrie, Orion Weller, James Mayfield
- Participant: hltcoe
- Paper: https://trec.nist.gov/pubs/trec33/papers/hltcoe.neuclir.pdf
- Runs: fas-hltcoe-MNED-plaid_distill_engeng_zs2engfas | fas-hltcoe-MNED-plaid_distill_engeng_zs2fasfas | fas-hltcoe-MNED-plaid_distill_engfas | fas-hltcoe-MNED-plaid_distill_engfas.mt5rerank.gpt4rerank | fas-hltcoe-MNED-plaid_distill_engfas_450p | fas-hltcoe-MNED-plaid_distill_engfas_termpool2 | fas-hltcoe-MNED-plaid_distill_engmlir | fas-hltcoe-MNED-plaid_distill_engfas.mt5rerank | fas-hltcoe-MNED-plaid_distill_mono_fasfas | fas-hltcoe-MNED-plaid_eqsynms_distill_engfas | fas-hltcoe-MNED-plaid_syn_distill_engfas | fas-hltcoe-MNTEH-kitchen_rankfuse | fas-hltcoe-MNTEH-kitchen_rankfuse,mt5rerank | fas-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.gpt4rerank | fas-hltcoe-MTED-plaid_distill_engeng | rus-hltcoe-MNED-plaid_distill_engeng_zs2engrus | rus-hltcoe-MNED-plaid_distill_engeng_zs2rusrus | rus-hltcoe-MNED-plaid_distill_engmlir | rus-hltcoe-MNED-plaid_distill_engrus | rus-hltcoe-MNED-plaid_distill_engrus.mt5rerank | rus-hltcoe-MNED-plaid_distill_engrus.mt5rerank.gpt4rerank | rus-hltcoe-MNED-plaid_distill_engrus_450p | rus-hltcoe-MNED-plaid_distill_engrus_termpool2 | rus-hltcoe-MNED-plaid_distill_mono_rusrus | rus-hltcoe-MNED-plaid_eqsynms_distill_engrus | rus-hltcoe-MNED-plaid_syn_distill_engrus | rus-hltcoe-MNTEH-kitchen_rankfuse | rus-hltcoe-MNTEH-kitchen_rankfuse,mt5rerank | rus-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.gpt4rerank | rus-hltcoe-MTED-plaid_distill_engeng | zho-hltcoe-MNED-plaid_distill_engeng_zs2engzho | zho-hltcoe-MNED-plaid_distill_engeng_zs2zhozho | zho-hltcoe-MNED-plaid_distill_engmlir | zho-hltcoe-MNED-plaid_distill_engzho | zho-hltcoe-MNED-plaid_distill_engzho.mt5rerank | zho-hltcoe-MNED-plaid_distill_engzho.mt5rerank.gpt4rerank | zho-hltcoe-MNED-plaid_distill_engzho_450p | zho-hltcoe-MNED-plaid_distill_engzho_termpool2 | zho-hltcoe-MNED-plaid_distill_mono_zhozho | zho-hltcoe-MNED-plaid_eqsynms_distill_engzho | zho-hltcoe-MNED-plaid_syn_distill_engzho | zho-hltcoe-MNTEH-kitchen_rankfuse | zho-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank | zho-hltcoe-MTED-plaid_distill_engeng | zho-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.gpt4rerank | tech-hltcoe-MNED-plaid_distill_engeng_zs2engzho | tech-hltcoe-MNED-plaid_distill_engeng_zs2zhozho | tech-hltcoe-MNED-plaid_distill_engzho | tech-hltcoe-MNED-plaid_distill_zhozho | tech-hltcoe-MNED-plaid_syn_distill_engzho | tech-hltcoe-MNTEH-kitchen_rankfuse | tech-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank | tech-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.gpt4rerank | tech-hltcoe-MTED-plaid_distill_engeng | tech-hltcoe-MNED-plaid_eqsynms_distill_engzho | mlir-hltcoe-MNED-plaid_distill_clir_scorefuse | mlir-hltcoe-MNED-plaid_distill_clir.mt5rerank.scorefuse | mlir-hltcoe-MNED-plaid_distill_clir.mt5rerank.scorefuse.gpt4rerank | mlir-hltcoe-MNED-plaid_distill_engeng_zs | mlir-hltcoe-MNED-plaid_distill_mlir_bycoll_scorefuse | mlir-hltcoe-MNED-plaid_distill_mlir_mixedentry | mlir-hltcoe-MNED-plaid_distill_mlir_mixedentry_termpool2 | mlir-hltcoe-MNED-plaid_distill_mlir_mixedpass | mlir-hltcoe-MNED-plaid_distill_mlir_rr | mlir-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.scorefuse | mlir-hltcoe-MNTEH-kitchen_rankfuse.mt5rerank.scorefuse.gpt4rerank | mlir-hltcoe-MTED-plaid_distill_engeng | zho-jhu-orion-aggregated-w-gpt4o | rus-jhu-orion-aggregated-w-gpt4o | fas-jhu-orion-aggregated-w-gpt4o | zho-jhu-orion-aggregated-w-claude | rus-jhu-orion-aggregated-w-claude | fas-jhu-orion-aggregated-w-claude | fas-hltcoe-eugene-gpt4o | rus-hltcoe-eugene-gpt4o | fas-hltcoe-eugene-gpt35turbo | rus-hltcoe-eugene-gpt35turbo | zho-hltcoe-eugene-gpt35turbo | zho-hltcoe-eugene-gpt4o-fixed
Abstract
The HLTCOE team applied PLAID, an mT5 reranker, GPT-4 reranker, score fusion, and document translation to the TREC 2024 NeuCLIR track. For PLAID we included a variety of models and training techniques – Translate Distill (TD), Generate Distill (GD) and multi-lingual translate-distill (MTD). TD uses scores from the mT5 model over English MS MARCO query-document pairs to learn how to score query-document pairs where the documents are translated to match the CLIR setting. GD follows TD but uses passages from the collection and queries generated by an LLM for training examples. MTD uses MS MARCO translated into multiple languages, allowing experiments on how to batch the data during training. Finally, for report generation we experimented with system combination over different runs. One family of systems used either GPT-4o or Claude-3.5-Sonnet to summarize the retrieved results from a series of decomposed sub-questions. Another system took the output from those two models and verified/combined them with Claude-3.5-Sonnet. The other family used GPT4o and GPT3.5Turbo to extract and group relevant facts from the retrieved documents based on the decomposed queries. The resulting submissions directly concatenate the grouped facts to form the report and their documents of origin as the citations. The team submitted runs to all NeuCLIR tasks: CLIR and MLIR news tasks as well as the technical documents task and the report generation task.
Bibtex
@inproceedings{hltcoe-trec2024-papers-proc-1,
title = {HLTCOE at TREC 2024 NeuCLIR Track},
author = {Eugene Yang and Dawn Lawrie and Orion Weller and James Mayfield},
booktitle = {Proceedings of the 33th Text {REtrieval} Conference (TREC 2024)},
year = {2024},
address = {Gaithersburg, Maryland},
series = {NIST SP 1329}
}