Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
PythonTIGER-AI-Lab/VLM2Vec

VLM2Vec

This repo contains the code for "VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks" [ICLR 2025]

70.9/100
586Forks: 51
View on GitHubHomepage →
Loading report...

Similar Projects

all-in-rag

68

🔍大模型应用开发实战一:RAG 技术全栈指南,在线阅读地址:https://datawhalechina.github.io/all-in-rag/

Python4.4K

ms-swift

88

Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs (Qwen3.5, DeepSeek-R1, GLM-5, InternLM3, Llama4, ...) and 300+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, GLM4.5v, Llava, Phi4, ...) (AAAI 2025).

Python13.0K

VLM-R1

55

Solve Visual Understanding with Reinforced VLMs

Python5.9K

UltraRAG

85

A Low-Code MCP Framework for Building Complex and Innovative RAG Pipelines

Python5.4K
Back to List