Posts
-
InternVL2-MPO: Enhancing the Reasoning Ability of Multimodal Large Language Models via Mixed Preference Optimization
Mini-InternVL 2.0: A Flexible-Transfer Pocket Multimodal Model with 5% Parameters and 90% Performance
Mono-InternVL: Pushing the Boundaries of Monolithic Multimodal Large Language Models with Endogenous Visual Pre-training
InternOmni: Extending InternVL with Audio Modality
InternVL2: Better than the Best—Expanding Performance Boundaries of Open-Source Multimodal Models with the Progressive Scaling Strategy
ShareGPT-4o: Comprehensive Multimodal Annotations With GPT-4o
Mini-InternVL 1.5: A Powerful Pocket Multimodal Model with 8% Parameters for 80% Performance
InternVL 1.5: How Far Are We to GPT-4V? Closing the Gap to Commercial Multimodal Models with Open-Source Suites
InternVL 1.2: Scaling up LLM to 34B
InternVL 1.1: Enhance Chinese and OCR Capabilities
InternVL 1.0: Scaling up Vision Foundation Models and Aligning for Generic Visual-Linguistic Tasks