The implementation for SIGIR 2026: Learning to Retrieve from Agent Trajectories.
README
[SIGIR 2026] Learning to Retrieve from Agent Trajectories
Retrieval is no longer optimized only for human searchers. As large language model agents increasingly issue queries, inspect snippets, browse documents, and reason over retrieved evidence, the target of retrieval training has shifted from human interaction to agent interaction. LRAT studies this paradigm shift and learns retrievers directly from multi-step agent trajectories.
Training data for agent native search should match how search agents actually search, browse, and consume evidence.
LRAT studies how to train retrievers from the intermediate behaviors of strong search agents rather than from only final answers. The repository focuses on a practical pipeline for:
collecting long-horizon search trajectories from agentic systems,
converting trajectories into retrieval supervision,
training retrievers on the resulting samples, and
evaluating both retrieval quality and end-to-end task success.
Trajectory-first retrieval learning: build retriever supervision from agent search and browse traces instead of relying only on static relevance labels.
Agent-friendly data collection: run local or API-based research agents and save each query as structured trajectory JSON.
Training data construction with an LLM judge: turn trajectories into (query, pos, neg, ...) training pairs with reasoning-aware annotations.
Benchmark-oriented evaluation: evaluate outputs on BrowseComp-Plus and InfoSeek-Eval with a local vLLM judge.
Core utilities for index construction and trajectory-to-training-data conversion
search_agent/
Agent clients for Tongyi DeepResearch, WebExplorer, AgentCPM, OpenAI-compatible APIs, and related prompts/utilities
searcher/
Search backends and local retrieval interfaces
docs/
Step-by-step documentation for indexing, trajectory construction, training data construction, and evaluation
datasets/
Benchmark files used in evaluation
topics-qrels/
Query and qrel files for retrieval experiments
trajectory/
Example trajectory artifacts
FlagEmbedding/
Local copy of FlagEmbedding used for retriever training
tevatron/
Local copy of Tevatron utilities used in dense retrieval workflows
scripts_evaluation/
Evaluation scripts for end-to-end judging
Vendored Dependencies
FlagEmbedding/ is a vendored and locally modified copy based on the upstream FlagEmbedding project. In this repository, it reflects user-side modifications layered on top of upstream work and earlier external changes.
tevatron/ is a vendored upstream dependency used to support dense retrieval utilities and encoding workflows.
If you do not want to build training data from scratch, you can directly use the released LRAT-Train dataset. If you prefer to control filtering or supervision design yourself, you can also start from saved agent trajectories and rerun pair extraction with src/data_builder.py.
You can plug the JSONL generated by src/data_builder.py into your existing training setup without changing the repository-level presentation structure.
This repository is released under the Apache License 2.0. See LICENSE.
Vendored components keep their own upstream licenses, especially:
FlagEmbedding/ under its upstream MIT license
tevatron/ under Apache License 2.0
Citation
If you find this repository useful, please cite our SIGIR 2026 paper below. The latest public version is available on arXiv.
@inproceedings{zhou2026lrat,
title={Learning to Retrieve from Agent Trajectories},
author={Zhou, Yuqi and Dai, Sunhao and Qu, Changle and Pang, Liang and Xu, Jun and Wen, Ji-Rong},
booktitle={Proceedings of the 49th International ACM SIGIR Conference on Research and Development in Information Retrieval},
year={2026}
}
Release History
Version
Changes
Urgency
Date
0.0.0
No release found โ using repo HEAD
High
4/8/2026
main@2026-04-08
Latest activity on main branch
High
4/8/2026
main@2026-04-08
Latest activity on main branch
High
4/8/2026
main@2026-04-08
Latest activity on main branch
High
4/8/2026
main@2026-04-08
Latest activity on main branch
High
4/8/2026
main@2026-04-08
Latest activity on main branch
Medium
4/8/2026
main@2026-04-08
Latest activity on main branch
Medium
4/8/2026
main@2026-04-08
Latest activity on main branch
Medium
4/8/2026
main@2026-04-08
Latest activity on main branch
Medium
4/8/2026
main@2026-04-08
Latest activity on main branch
Medium
4/8/2026
Dependencies & License Audit
Loading dependencies...
Similar Packages
adk-pythonAn open-source, code-first Python toolkit for building, evaluating, and deploying sophisticated AI agents with flexibility and control.v1.31.1
OpenOutreachLinkedin Automation Tool: Describe your product. Define your target market. The AI finds the leads for you.main@2026-04-20
AReaLLightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible.v1.0.3
MemOSAI memory OS for LLM and Agent systems(moltbot,clawdbot,openclaw), enabling persistent Skill memory for cross-task skill reuse and evolution.v2.0.13
taijiAI-powered self-learning OS with I Ching philosophy | ่ๅๆ็ปๅฒๅญฆ็่ชๅญฆๅ AI ๆไฝ็ณป็ปv0.1.0