EvoGround: Self-Evolving Video Agents for Video Temporal Grounding

1Seoul National University,  2Northeastern University

Abstract

Video temporal grounding (VTG) takes an untrimmed video and a natural language query as input and localizes the temporal moment that best matches the query. While remarkable progress has been made, existing methods still rely on large-scale, task-specific datasets that require costly and labor-intensive manual annotation. In this paper, we introduce EvoGround, a framework of two coupled self-evolving agents, a proposer and a solver, that learn temporal information from raw videos without any human-labeled data. The proposer generates query–moment pairs from raw videos, while the solver learns to ground them and provides feedback that improves the proposer in return. Through this self-reinforcing loop, driven entirely by reinforcement learning, the two agents mutually improve each other across iterations. Despite never seeing any training data, EvoGround matches or outperforms most fully supervised models across multiple VTG benchmarks, while also demonstrating strong fine-grained video captioning capability.

Overview

The Self-evolution Loop of EvoGround.

The Self-evolution Loop of EvoGround. The proposer generates query (q)–moment (m) pairs from a given video, and the solver learns from these pairs and provides feedback signals through its moment predictions (). As the process is repeated, the two agents mutually improve each other.

Method

Overview of EvoGround.

EvoGround consists of two agents: a proposer and a solver. The proposer identifies candidate temporal events from raw videos and generates corresponding query–moment pairs, while the solver learns to ground temporal moments using the generated data. Both agents are initialized from the same backbone and evolve solely through a self-reinforcing loop without any labeled data.

The proposer is guided by three reward criteria: validity (format reward), consistency (consistency reward), and solvability (feedback reward). The consistency reward measures intra-consistency — how coherently a query aligns with frames within its moment — and inter-consistency — how discriminatively the query matches its own moment relative to others. The feedback reward uses the solver's accuracy (measured by timestamp-aware IoU) as a signal of whether generated pairs are learnable.

The solver is trained on proposer-generated query–moment pairs using a format reward and an accuracy reward (tIoU). EvoGround adopts GDPO as its RL optimizer and a curriculum design that progressively increases the solvability threshold across iterations, shifting focus from coarse matches toward more precise temporal alignments.

Results

Despite never seeing any manual annotations during training, EvoGround demonstrates strong performance across all benchmarks. It surpasses all SFT-based models and achieves first- or second-best performance among RL-based models — which rely on manually labeled data.

Charades-STA & ActivityNet-Captions

Method Charades-STA ActivityNet-Captions
R1@0.3R1@0.5R1@0.7mIoU R1@0.3R1@0.5R1@0.7mIoU
SFT-based Models
TimeChat-32.213.432.236.220.29.521.8
VTimeLLM51.027.511.431.244.027.814.330.4
Qwen2.5-VL68.548.822.545.038.725.614.928.6
TimeSuite69.948.724.0-----
VideoChat-Flash74.553.127.6-----
RL-based Models
Time-R178.160.835.358.158.639.021.440.5
EvoGround (Ours)76.458.633.652.561.642.525.042.8

TVGBench, ReXTime & E.T.Bench

Method TVGBench ReXTime E.T.Bench
R1@0.3R1@0.5R1@0.7mIoU R1@0.3R1@0.5mIoU TVG F1
SFT-based Models
TimeChat22.411.95.3-14.47.611.626.2
Qwen2.5-VL28.119.510.520.416.311.213.346.6
TRACE37.025.514.6-----
RL-based Models
VideoChat-R1.531.520.811.322.028.318.020.550.3
Time-R139.328.016.027.832.222.124.169.4
EvoGround (Ours)42.128.515.329.433.522.425.569.8

On TemporalBench (fine-grained video captioning), EvoGround outperforms all prior models across every metric — including CIDEr and BLEU — despite never being explicitly trained for captioning. This suggests that the self-evolution loop fosters a broader and more grounded understanding of video content.

Analysis

Reward dynamics across iterations. As the proposer evolves over iterations, the solver correspondingly demonstrates progressively higher accuracy.

Reward dynamics across iterations.

Improvements across iterations. GDPO consistently outperforms GRPO. Increasing the solvability threshold δ improves performance on longer videos and moments.

Performance improvements across iterations.

Data distributions across reward configurations. Each reward shapes the generated data differently — the consistency reward produces tighter moments, while the feedback reward enriches query descriptiveness.

Data distributions under different reward configurations.

Qualitative Results

Qualitative captioning example.

EvoGround generates detailed, temporally-grounded descriptions of video content. The proposer produces event-level captions that accurately capture the sequence of actions and participants, closely matching the ground-truth despite receiving no caption supervision.

BibTeX

@article{jung2025evoground,
  title     = {EvoGround: Self-Evolving Video Agents for Video Temporal Grounding},
  author    = {Jung, Minjoon and Zhang, Byoung-Tak and Torresani, Lorenzo},
  journal   = {arXiv preprint},
  year      = {2025}
}