Update stage2-a-llava-qwen/config.json
Browse files
stage2-a-llava-qwen/config.json
CHANGED
|
@@ -1,5 +1,4 @@
|
|
| 1 |
{
|
| 2 |
-
"_name_or_path": "/share/project/gushuhao/external_model/Qwen2___5-1___5B-Instruct",
|
| 3 |
"architectures": [
|
| 4 |
"LlavaQwenForCausalLM"
|
| 5 |
],
|
|
@@ -45,7 +44,7 @@
|
|
| 45 |
"mm_use_im_start_end": false,
|
| 46 |
"mm_vision_select_feature": "patch",
|
| 47 |
"mm_vision_select_layer": -2,
|
| 48 |
-
"mm_vision_tower": "/
|
| 49 |
"mm_vision_tower_lr": 2e-06,
|
| 50 |
"model_type": "qwen2",
|
| 51 |
"num_attention_heads": 12,
|
|
|
|
| 1 |
{
|
|
|
|
| 2 |
"architectures": [
|
| 3 |
"LlavaQwenForCausalLM"
|
| 4 |
],
|
|
|
|
| 44 |
"mm_use_im_start_end": false,
|
| 45 |
"mm_vision_select_feature": "patch",
|
| 46 |
"mm_vision_select_layer": -2,
|
| 47 |
+
"mm_vision_tower": "google/siglip-so400m-patch14-384",
|
| 48 |
"mm_vision_tower_lr": 2e-06,
|
| 49 |
"model_type": "qwen2",
|
| 50 |
"num_attention_heads": 12,
|