Grounded Retrieval and Agentic Intelligence for Vision-Language plan > retrieve > reason > verify
A CVPR 2026 workshop for researchers and practitioners building grounded multimodal retrieval, reranking, and verification systems that can be deployed with confidence.
P-R-R-V Loop
Plan tasks and tool use, retrieve candidates, rerank with grounded reasoning, then verify with calibrated evidence.
Focus areas at a glance
We’re looking for work that ties mutlimodal agentic tools,evidence to decisions, scales retrieval and reranking, and evaluates real deployment constraints.
Secondary Track
Visual Grounding & Evidence
Retrieval & Ranking
Agentic Tools & Planning
Evaluation & Efficiency
Why now
Grounded evidence is the missing piece for agentic vision
Vision-language agents increasingly rely on a loop of plan, retrieve, rerank, and verify before acting. But how we measure evidence grounding, calibration, and end-to-end efficiency is still fragmented across communities.
GRAIL-V brings together CV, IR, NLP, HCI, and systems researchers and practitioners working on evidence-centric retrieval and verification for deployable agentic systems.
Why submit or attend
- Get feedback on grounded retrieval, reranking, and verification.
- Share real-world evaluations, demos, and deployment lessons.
- Connect with researchers building agentic VLM pipelines.
- Help shape community benchmarks and best practices.
Submission details
We invite archival papers and demos on grounded multimodal retrieval, reranking, and verification for agentic vision systems. We especially welcome work that reports grounded evidence (region/page/moment), calibration or abstention behavior, and real deployment constraints (latency, memory, cost).
Submission at a glance
- CVPR format, double-blind, up to 8 pages (references excluded).
- Grounded evidence + efficiency metrics strongly encouraged.
- Submit via OpenReview.
Important dates
Timeline (Anywhere on Earth)
Paper submission deadline
OpenReview submission closes at 23:59 AoE.
Notification to authors
Decisions released via OpenReview.
Camera-ready due
Final versions for CVPR workshop proceedings.
Workshop day in Denver, USA
Full-day program with keynotes, panels, and posters.
Updates
News and deadlines
GRAIL-V has been accepted as a CVPR 2026 workshop in Denver, USA.
Submit your paper via the GRAIL-V OpenReview forum.
Speakers
Kristen Grauman
University of Texas at Austin
Leads work on egocentric video understanding and retrieval at scale, with a focus on long-horizon grounding and efficient perception.Mohit Bansal
University of North Carolina at Chapel Hill
Researches multimodal language/vision agents, grounded reasoning, and controllable generation for real-world tasks.Dan Roth
University of Pennsylvania / Oracle AI
Pioneer in structured, grounded reasoning and robust inference for language/vision systems deployed in real settings.In-person with hybrid support
Location
Denver, USA. Workshop dates: Jun 3-4, 2026.
Hybrid plan
Hybrid participation follows CVPR guidance; details will be posted when available.
Registration
Registration is handled via CVPR. We will link the official registration page when it opens.
Accessibility
Reserved front-row seating, mic runners, captioning support.
Contact
Stay connected
General inquiries
Reach out to the organizers with questions about submissions, sponsorship, or program.