15 Latest Papers: Video & Multimodal Retrieval (August 2025)

by Felix Dubois 61 views

Hey everyone! It's August 11, 2025, and we've got a fresh batch of cutting-edge research papers in the realms of video and multimodal retrieval. This is your go-to spot for staying updated on the latest advancements in these rapidly evolving fields. We'll be breaking down the key highlights from these papers, making it easy for you to understand the core ideas and their potential impact. So, let's dive in and explore the fascinating world of video and multimodal retrieval!

Please check the Github page for a better reading experience and more papers. You guys know the drill – the GitHub page is always the best place for the full experience, with all the bells and whistles.

Video Retrieval

Let's kick things off with video retrieval. This area is all about efficiently finding the videos you need from vast datasets. From enhancing search accuracy to tackling biases in AI-generated content, there's a lot happening. We'll explore how researchers are leveraging multimodal large language models, addressing cold-start challenges in short-form video recommendations, and even quantifying uncertainty in interactive retrieval systems. Get ready for a journey through the innovations shaping the future of video search!

Title Date Comment
Bidirectional Likelihood Estimation with Multi-Modal Large Language Models for Text-Video Retrieval 2025-08-06 ICCV 2025 Highlight
GAID: Frame-Level Gated Audio-Visual Integration with Directional Perturbation for Text-Video Retrieval 2025-08-03
Generative Ghost: Investigating Ranking Bias Hidden in AI-Generated Videos 2025-07-29
13 pa...

13 pages, Accepted at ACMMM2025

T2VParser: Adaptive Decomposition Tokens for Partial Alignment in Text to Video Retrieval 2025-07-28
HLFormer: Enhancing Partially Relevant Video Retrieval with Hyperbolic Learning 2025-07-27
Accep...

Accepted by ICCV'25. 13 pages, 6 figures, 4 tables

Short-Form Video Recommendations with Multimodal Embeddings: Addressing Cold-Start and Bias Challenges 2025-07-25
FrameFusion: Combining Similarity and Importance for Video Token Reduction on Large Vision Language Models 2025-07-24 ICCV 2025
Quantifying and Narrowing the Unknown: Interactive Text-to-Video Retrieval via Uncertainty Minimization 2025-07-24
Accep...

Accepted by ICCV 2025

Prompt-aware of Frame Sampling for Efficient Text-Video Retrieval 2025-07-21
U-MARVEL: Unveiling Key Factors for Universal Multimodal Retrieval via Embedding Learning with MLLMs 2025-07-20
Techn...

Technical Report (in progress)

Smart Routing for Multimodal Video Retrieval: When to Search What 2025-07-12
Accep...

Accepted to ICCV 2025 Multimodal Representation and Retrieval Workshop

MSVD-Indonesian: A Benchmark for Multimodal Video-Text Tasks in Indonesian 2025-07-12
10 pa...

10 pages, 5 figures, 5 tables

BiMa: Towards Biases Mitigation for Text-Video Retrieval via Scene Element Guidance 2025-07-07
Accep...

Accepted at ACM MM 2025

VLM2Vec-V2: Advancing Multimodal Embedding for Videos, Images, and Visual Documents 2025-07-07 Technical Report
Are Synthetic Videos Useful? A Benchmark for Retrieval-Centric Evaluation of Synthetic Videos 2025-07-03 7 pages, 10 figures

Key Papers in Video Retrieval: A Closer Look

Let's zoom in on some of the standout papers in video retrieval. First up is "Bidirectional Likelihood Estimation with Multi-Modal Large Language Models for Text-Video Retrieval," an ICCV 2025 Highlight. This paper explores how multimodal large language models (MLLMs) can be used to improve text-video retrieval by leveraging bidirectional likelihood estimation. The core idea is to use the MLLM to understand the relationship between text and video, enhancing the accuracy of retrieval systems. This is crucial for creating systems that can truly understand the content of a video, not just match keywords.

Next, we have "GAID: Frame-Level Gated Audio-Visual Integration with Directional Perturbation for Text-Video Retrieval." This research delves into integrating audio and visual information at the frame level. By using gated audio-visual integration and directional perturbation, the model becomes more robust and accurate in retrieving videos based on textual queries. Think about how much richer a video understanding system can be when it processes both what you see and what you hear!

Another fascinating paper is "Generative Ghost: Investigating Ranking Bias Hidden in AI-Generated Videos." This work addresses a critical issue: the biases that can creep into AI-generated videos. By understanding these biases, we can develop methods to mitigate them, ensuring fairer and more accurate retrieval results. Imagine a search engine that favors certain types of content simply because of its AI training – this research aims to prevent that.

"T2VParser: Adaptive Decomposition Tokens for Partial Alignment in Text to Video Retrieval" presents an adaptive approach to handling partial alignment between text and video. The method uses decomposition tokens to better match textual descriptions with video segments, making retrieval more precise. This is particularly useful for complex videos where the description might not perfectly align with every frame.

"HLFormer: Enhancing Partially Relevant Video Retrieval with Hyperbolic Learning" explores the use of hyperbolic learning to improve retrieval performance. By embedding video and text in a hyperbolic space, the model can better capture the hierarchical relationships between concepts, leading to enhanced retrieval accuracy. This is a clever way to represent complex data in a way that reflects its underlying structure.

Finally, "FrameFusion: Combining Similarity and Importance for Video Token Reduction on Large Vision Language Models" introduces a method for reducing video tokens in large vision language models (LVLMs). By combining similarity and importance measures, the model can focus on the most relevant parts of a video, improving efficiency without sacrificing accuracy. This is a crucial step towards making LVLMs more practical for real-world applications.

These papers highlight the diverse approaches being taken in video retrieval research, from leveraging MLLMs and audio-visual integration to addressing biases and improving efficiency. The field is clearly moving towards more nuanced and accurate video understanding, which will have a huge impact on how we search for and interact with video content.

Multimodal Retrieval

Now, let's shift our focus to multimodal retrieval. This field takes things a step further by incorporating multiple types of data – think text, images, audio, and more – to enhance retrieval accuracy. We'll be looking at papers that explore everything from knowledge graph-enhanced retrieval for visual question answering to using multimodal retrieval for understanding protein function. This is where the real magic happens, as we unlock the power of combining different data modalities!

Title Date Comment
M2IO-R1: An Efficient RL-Enhanced Reasoning Framework for Multimodal Retrieval Augmented Multimodal Generation 2025-08-08
Explaining Similarity in Vision-Language Encoders with Weighted Banzhaf Interactions 2025-08-07 Preprint
mKG-RAG: Multimodal Knowledge Graph-Enhanced RAG for Visual Question Answering 2025-08-07
UniFGVC: Universal Training-Free Few-Shot Fine-Grained Vision Classification via Attribute-Aware Multimodal Retrieval 2025-08-06
Understanding protein function with a multimodal retrieval-augmented foundation model 2025-08-05
ArtSeek: Deep artwork understanding via multimodal in-context reasoning and late interaction retrieval 2025-07-29
PUMA: Layer-Pruned Language Model for Efficient Unified Multimodal Retrieval with Modality-Adaptive Learning 2025-07-28
Accep...

Accepted to ACM MM 2025

VL-CLIP: Enhancing Multimodal Recommendations via Visual Grounding and LLM-Augmented CLIP Embeddings 2025-07-22
Accep...

Accepted at RecSys 2025; DOI:https://doi.org/10.1145/3705328.3748064

U-MARVEL: Unveiling Key Factors for Universal Multimodal Retrieval via Embedding Learning with MLLMs 2025-07-20
Techn...

Technical Report (in progress)

Evaluating Multimodal Large Language Models on Educational Textbook Question Answering 2025-07-15 8 Pages
DeepWriter: A Fact-Grounded Multimodal Writing Assistant Based On Offline Knowledge Base 2025-07-14 work in process
Smart Routing for Multimodal Video Retrieval: When to Search What 2025-07-12
Accep...

Accepted to ICCV 2025 Multimodal Representation and Retrieval Workshop

Llama Nemoretriever Colembed: Top-Performing Text-Image Retrieval Model 2025-07-07
MOTOR: Multimodal Optimal Transport via Grounded Retrieval in Medical Visual Question Answering 2025-06-28
Universal Retrieval for Multimodal Trajectory Modeling 2025-06-27
18 pa...

18 pages, 3 figures, accepted by Workshop on Computer-use Agents @ ICML 2025

Diving Deeper into Multimodal Retrieval Papers

Let's shine a spotlight on some key papers in the multimodal retrieval space. "M2IO-R1: An Efficient RL-Enhanced Reasoning Framework for Multimodal Retrieval Augmented Multimodal Generation" introduces a novel framework that uses reinforcement learning (RL) to enhance reasoning in multimodal systems. By combining retrieval-augmented generation with RL, the model can generate more coherent and relevant responses. This is a big step towards creating AI that can not only understand but also generate rich multimodal content.

"mKG-RAG: Multimodal Knowledge Graph-Enhanced RAG for Visual Question Answering" explores the use of multimodal knowledge graphs to improve visual question answering (VQA). By integrating knowledge graphs, the model can access a wealth of structured information, leading to more accurate and insightful answers. This is like giving the AI a super-powered brain filled with interconnected facts!

"Understanding protein function with a multimodal retrieval-augmented foundation model" tackles a complex problem in bioinformatics: predicting protein function. By using a multimodal retrieval-augmented foundation model, the researchers can leverage diverse data sources to better understand the intricate workings of proteins. This has huge implications for drug discovery and personalized medicine.

"ArtSeek: Deep artwork understanding via multimodal in-context reasoning and late interaction retrieval" dives into the world of art, using multimodal techniques to understand and appreciate artwork. By combining visual and textual information, the model can reason about the context and meaning of art, providing a deeper level of understanding. This is a fascinating application of AI that bridges the gap between technology and the humanities.

"PUMA: Layer-Pruned Language Model for Efficient Unified Multimodal Retrieval with Modality-Adaptive Learning" presents a method for creating efficient multimodal retrieval systems by pruning layers in a language model. This technique, combined with modality-adaptive learning, allows the model to focus on the most relevant information from different modalities, improving both speed and accuracy. Efficiency is key when dealing with massive datasets, so this research is crucial.

"VL-CLIP: Enhancing Multimodal Recommendations via Visual Grounding and LLM-Augmented CLIP Embeddings" focuses on enhancing multimodal recommendations by using visual grounding and large language model (LLM)-augmented CLIP embeddings. This approach allows the model to better understand the visual content of items, leading to more personalized and relevant recommendations. Imagine a recommendation system that truly gets your visual preferences – that's the goal here.

These papers showcase the breadth and depth of research in multimodal retrieval, from enhancing question answering and understanding protein function to appreciating art and improving recommendation systems. By combining different data modalities, we're unlocking new possibilities for AI to understand and interact with the world around us.

Conclusion

Alright, guys, that's a wrap for this edition of the latest research papers! We've explored some fascinating advancements in both video and multimodal retrieval. From leveraging large language models to mitigating biases in AI-generated content, the field is buzzing with innovation. Whether you're a seasoned researcher or just getting started, I hope this overview has given you a valuable glimpse into the future of retrieval technologies. Stay tuned for more updates, and don't forget to check out the GitHub page for the full details on these papers. Keep exploring, keep innovating, and I'll catch you in the next one!