-
UniME-V2: MLLM-as-a-Judge for Universal Multimodal Embedding Learning
Paper • 2510.13515 • Published • 11 -
TianchengGu/UniME-V2-LLaVA-OneVision-8B
Image-Text-to-Text • 8B • Updated • 11 • 2 -
TianchengGu/UniME-V2-Qwen2VL-7B
Image-Text-to-Text • 8B • Updated • 912 • 2 -
TianchengGu/UniME-V2-Qwen2VL-2B
Image-Text-to-Text • 2B • Updated • 1.14k • 2
Kaicheng Yang
Kaichengalex
AI & ML interests
Multimodal Representation Learning/ Vision-Language Pretraining/DeepResearch
Recent Activity
liked
a model
2 days ago
zai-org/GLM-4.7
upvoted
a
paper
7 days ago
HyperVL: An Efficient and Dynamic Multimodal Large Language Model for Edge Devices
upvoted
a
paper
8 days ago
Towards Scalable Pre-training of Visual Tokenizers for Generation