imp-v1-3b
I
Imp V1 3b
Overview :
The Imp project aims to provide a series of powerful multimodal small language models (MSLMs). Our imp-v1-3b is a powerful 3-billion parameter MLM built upon a small but powerful SLM Phi-2 (2.7 billion) and a powerful visual encoder SigLIP (400 million), trained on the LLaVA-v1.5 training dataset. Imp-v1-3b significantly outperforms similar-sized models on various multimodal benchmark tests, even showing slight superiority over the powerful LLaVA-7B model on some multimodal benchmarks.
Target Users :
Used for multimodal tasks such as natural language processing and visual question answering.
Total Visits: 29.7M
Top Region: US(17.94%)
Website Views : 53.3K
Use Cases
Used for natural language processing tasks
Used for visual question answering tasks
Used for multimodal tasks
Features
Provides a powerful multimodal small language model
Has 3 billion parameters
Exhibits excellent performance on multimodal benchmark tests
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase