Update README.md
Browse files
README.md
CHANGED
|
@@ -12,6 +12,7 @@ tags:
|
|
| 12 |
- vision
|
| 13 |
- ocr
|
| 14 |
- custom_code
|
|
|
|
| 15 |
---
|
| 16 |
|
| 17 |
# Mono-InternVL-2B
|
|
@@ -30,7 +31,7 @@ tags:
|
|
| 30 |
|
| 31 |
## Introduction
|
| 32 |
|
| 33 |
-
We release Mono-InternVL, a **monolithic** multimodal large language model (MLLM) that integrates visual encoding and textual decoding into a single LLM. In Mono-InternVL, a set of visual experts is embedded into the pre-trained LLM via a mixture-of-experts mechanism. By freezing the LLM, Mono-InternVL ensures that visual capabilities are optimized without compromising the pre-trained language knowledge. Based on this structure, an innovative Endogenous Visual Pretraining (EViP) is introduced to realize coarse-to-fine visual learning.
|
| 34 |
|
| 35 |
|
| 36 |
|
|
@@ -38,7 +39,7 @@ Mono-InternVL achieves superior performance compared to state-of-the-art MLLM Mi
|
|
| 38 |
|
| 39 |
|
| 40 |
|
| 41 |
-
This repository contains the instruction-tuned Mono-InternVL-2B model. It is built upon [internlm2-chat-1_8b](https://huggingface.co/internlm/internlm2-chat-1_8b). For more details, please refer to our [paper](https://arxiv.org/abs/2410.08202).
|
| 42 |
|
| 43 |
|
| 44 |
|
|
|
|
| 12 |
- vision
|
| 13 |
- ocr
|
| 14 |
- custom_code
|
| 15 |
+
- moe
|
| 16 |
---
|
| 17 |
|
| 18 |
# Mono-InternVL-2B
|
|
|
|
| 31 |
|
| 32 |
## Introduction
|
| 33 |
|
| 34 |
+
We release Mono-InternVL, a **monolithic** multimodal large language model (MLLM) that integrates visual encoding and textual decoding into a single LLM. In Mono-InternVL, a set of visual experts is embedded into the pre-trained LLM via a mixture-of-experts (MoE) mechanism. By freezing the LLM, Mono-InternVL ensures that visual capabilities are optimized without compromising the pre-trained language knowledge. Based on this structure, an innovative Endogenous Visual Pretraining (EViP) is introduced to realize coarse-to-fine visual learning.
|
| 35 |
|
| 36 |
|
| 37 |
|
|
|
|
| 39 |
|
| 40 |
|
| 41 |
|
| 42 |
+
This repository contains the instruction-tuned Mono-InternVL-2B model, which has 1.8B activated parameters (3B in total). It is built upon [internlm2-chat-1_8b](https://huggingface.co/internlm/internlm2-chat-1_8b). For more details, please refer to our [paper](https://arxiv.org/abs/2410.08202).
|
| 43 |
|
| 44 |
|
| 45 |
|