wikeeyang/UniWorld-V1-NF4

#1
Files changed (49) hide show
  1. .gitattributes +0 -2
  2. models/UniWorld-V1/.gitattributes +0 -36
  3. models/UniWorld-V1/added_tokens.json +0 -24
  4. models/UniWorld-V1/chat_template.json +0 -3
  5. models/UniWorld-V1/config.json +0 -120
  6. models/UniWorld-V1/flux-redux-siglipv2-512.bin +0 -3
  7. models/UniWorld-V1/generation_config.json +0 -6
  8. models/UniWorld-V1/merges.txt +0 -0
  9. models/UniWorld-V1/model-00001-of-00003.safetensors +0 -3
  10. models/UniWorld-V1/model-00002-of-00003.safetensors +0 -3
  11. models/UniWorld-V1/model-00003-of-00003.safetensors +0 -3
  12. models/UniWorld-V1/model.safetensors.index.json +0 -0
  13. models/UniWorld-V1/preprocessor_config.json +0 -29
  14. models/UniWorld-V1/special_tokens_map.json +0 -31
  15. models/UniWorld-V1/task_head_final.pt +0 -3
  16. models/UniWorld-V1/tokenizer.json +0 -3
  17. models/UniWorld-V1/tokenizer_config.json +0 -213
  18. models/UniWorld-V1/vocab.json +0 -0
  19. models/flux/LICENSE.md +0 -82
  20. models/flux/model_index.json +0 -32
  21. models/flux/put_them_to_flux +0 -1
  22. models/flux/scheduler/scheduler_config.json +0 -11
  23. models/flux/text_encoder/config.json +0 -25
  24. models/flux/text_encoder/model.safetensors +0 -3
  25. models/flux/text_encoder_2/config.json +0 -22
  26. models/flux/text_encoder_2/model.safetensors.index.json +0 -226
  27. models/flux/text_encoder_2/t5xxl-unchained-f16.safetensors +0 -3
  28. models/flux/tokenizer/merges.txt +0 -0
  29. models/flux/tokenizer/special_tokens_map.json +0 -30
  30. models/flux/tokenizer/tokenizer_config.json +0 -30
  31. models/flux/tokenizer/vocab.json +0 -0
  32. models/flux/tokenizer_2/special_tokens_map.json +0 -125
  33. models/flux/tokenizer_2/spiece.model +0 -3
  34. models/flux/tokenizer_2/tokenizer.json +0 -0
  35. models/flux/tokenizer_2/tokenizer1.json +0 -0
  36. models/flux/tokenizer_2/tokenizer_config.json +0 -940
  37. models/flux/transformer/config.json +0 -18
  38. models/flux/transformer/diffusion_pytorch_model.safetensors.index.json +0 -0
  39. models/flux/vae/config.json +0 -38
  40. models/flux/vae/diffusion_pytorch_model.safetensors +0 -3
  41. models/siglip2/README.md +0 -102
  42. models/siglip2/config.json +0 -22
  43. models/siglip2/model.safetensors +0 -3
  44. models/siglip2/preprocessor_config.json +0 -24
  45. models/siglip2/put_them_to_siglip2 +0 -1
  46. models/siglip2/special_tokens_map.json +0 -34
  47. models/siglip2/tokenizer.json +0 -3
  48. models/siglip2/tokenizer.model +0 -3
  49. models/siglip2/tokenizer_config.json +0 -2020
.gitattributes CHANGED
@@ -33,5 +33,3 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
- models/UniWorld-V1/tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
- models/siglip2/tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
models/UniWorld-V1/.gitattributes DELETED
@@ -1,36 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
36
- tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/UniWorld-V1/added_tokens.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "</tool_call>": 151658,
3
- "<tool_call>": 151657,
4
- "<|box_end|>": 151649,
5
- "<|box_start|>": 151648,
6
- "<|endoftext|>": 151643,
7
- "<|file_sep|>": 151664,
8
- "<|fim_middle|>": 151660,
9
- "<|fim_pad|>": 151662,
10
- "<|fim_prefix|>": 151659,
11
- "<|fim_suffix|>": 151661,
12
- "<|im_end|>": 151645,
13
- "<|im_start|>": 151644,
14
- "<|image_pad|>": 151655,
15
- "<|object_ref_end|>": 151647,
16
- "<|object_ref_start|>": 151646,
17
- "<|quad_end|>": 151651,
18
- "<|quad_start|>": 151650,
19
- "<|repo_name|>": 151663,
20
- "<|video_pad|>": 151656,
21
- "<|vision_end|>": 151653,
22
- "<|vision_pad|>": 151654,
23
- "<|vision_start|>": 151652
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/UniWorld-V1/chat_template.json DELETED
@@ -1,3 +0,0 @@
1
- {
2
- "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
3
- }
 
 
 
 
models/UniWorld-V1/config.json DELETED
@@ -1,120 +0,0 @@
1
- {
2
- "architectures": [
3
- "UnivaQwen2p5VLForConditionalGeneration"
4
- ],
5
- "attention_dropout": 0.0,
6
- "bos_token_id": 151643,
7
- "decay": 0.999,
8
- "denoise_tower": {
9
- "_attn_implementation_autoset": true,
10
- "denoise_projector_type": "mlp2x_gelu",
11
- "denoiser_config": {
12
- "_class_name": "FluxTransformer2DModel",
13
- "_diffusers_version": "0.30.0.dev0",
14
- "_name_or_path": "flux",
15
- "attention_head_dim": 128,
16
- "guidance_embeds": true,
17
- "in_channels": 64,
18
- "joint_attention_dim": 4096,
19
- "num_attention_heads": 24,
20
- "num_layers": 19,
21
- "num_single_layers": 38,
22
- "patch_size": 1,
23
- "pooled_projection_dim": 768
24
- },
25
- "denoiser_type": "flux",
26
- "input_hidden_size": 3584,
27
- "model_type": "univa_denoise_tower",
28
- "output_hidden_size": 4096,
29
- "torch_dtype": "bfloat16",
30
- "vae_input_hidden_size": 64,
31
- "vae_projector_type": "mlp2x_gelu"
32
- },
33
- "eos_token_id": 151645,
34
- "hidden_act": "silu",
35
- "hidden_size": 3584,
36
- "image_begin_token_id": 151652,
37
- "image_end_token_id": 151653,
38
- "image_token_id": 151655,
39
- "image_token_length": null,
40
- "initializer_range": 0.02,
41
- "intermediate_size": 18944,
42
- "inv_gamma": 1.0,
43
- "max_position_embeddings": 128000,
44
- "max_window_layers": 28,
45
- "min_decay": 0.0,
46
- "model_type": "univa_qwen2p5vl",
47
- "num_attention_heads": 28,
48
- "num_hidden_layers": 28,
49
- "num_key_value_heads": 4,
50
- "optimization_step": 40000,
51
- "power": 0.6666666666666666,
52
- "quantization_config": {
53
- "_load_in_4bit": true,
54
- "_load_in_8bit": false,
55
- "bnb_4bit_compute_dtype": "bfloat16",
56
- "bnb_4bit_quant_storage": "uint8",
57
- "bnb_4bit_quant_type": "nf4",
58
- "bnb_4bit_use_double_quant": false,
59
- "llm_int8_enable_fp32_cpu_offload": false,
60
- "llm_int8_has_fp16_weight": false,
61
- "llm_int8_skip_modules": [
62
- "lm_head"
63
- ],
64
- "llm_int8_threshold": 6.0,
65
- "load_in_4bit": true,
66
- "load_in_8bit": false,
67
- "quant_method": "bitsandbytes"
68
- },
69
- "rms_norm_eps": 1e-06,
70
- "rope_scaling": {
71
- "mrope_section": [
72
- 16,
73
- 24,
74
- 24
75
- ],
76
- "rope_type": "default",
77
- "type": "default"
78
- },
79
- "rope_theta": 1000000.0,
80
- "shortcut_image_embeds": false,
81
- "shortcut_image_embeds_scale": 0.5,
82
- "shortcut_projector_type": null,
83
- "sliding_window": 32768,
84
- "tie_word_embeddings": false,
85
- "torch_dtype": "bfloat16",
86
- "transformers_version": "4.50.3",
87
- "update_after_step": 0,
88
- "use_cache": true,
89
- "use_ema_warmup": false,
90
- "use_sliding_window": false,
91
- "video_token_id": 151656,
92
- "vision_config": {
93
- "depth": 32,
94
- "fullatt_block_indexes": [
95
- 7,
96
- 15,
97
- 23,
98
- 31
99
- ],
100
- "hidden_act": "silu",
101
- "hidden_size": 1280,
102
- "in_channels": 3,
103
- "in_chans": 3,
104
- "intermediate_size": 3420,
105
- "model_type": "qwen2_5_vl",
106
- "num_heads": 16,
107
- "out_hidden_size": 3584,
108
- "patch_size": 14,
109
- "spatial_merge_size": 2,
110
- "spatial_patch_size": 14,
111
- "temporal_patch_size": 2,
112
- "tokens_per_second": 2,
113
- "torch_dtype": "bfloat16",
114
- "window_size": 112
115
- },
116
- "vision_end_token_id": 151653,
117
- "vision_start_token_id": 151652,
118
- "vision_token_id": 151654,
119
- "vocab_size": 152064
120
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/UniWorld-V1/flux-redux-siglipv2-512.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:20502187b641b387d3253acb1e6f86b11b24b693f7458c5e94763e1abff93d4b
3
- size 129009784
 
 
 
 
models/UniWorld-V1/generation_config.json DELETED
@@ -1,6 +0,0 @@
1
- {
2
- "_from_model_config": true,
3
- "bos_token_id": 151643,
4
- "eos_token_id": 151645,
5
- "transformers_version": "4.50.3"
6
- }
 
 
 
 
 
 
 
models/UniWorld-V1/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
models/UniWorld-V1/model-00001-of-00003.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:1a85f0ecd05a3e32f557ac82c909b25243479a56b7d92789a37c1a8c35db876a
3
- size 4975533694
 
 
 
 
models/UniWorld-V1/model-00002-of-00003.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba4b4e8677e7088a3df7606921fd8dd6c5740ed0d6f519309e655b47605018db
3
- size 4982436876
 
 
 
 
models/UniWorld-V1/model-00003-of-00003.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f52e29b74459879ba9048c48056a742ade917f5c780bf1a23c6acf8a91848893
3
- size 3072962803
 
 
 
 
models/UniWorld-V1/model.safetensors.index.json DELETED
The diff for this file is too large to render. See raw diff
 
models/UniWorld-V1/preprocessor_config.json DELETED
@@ -1,29 +0,0 @@
1
- {
2
- "do_convert_rgb": true,
3
- "do_normalize": true,
4
- "do_rescale": true,
5
- "do_resize": true,
6
- "image_mean": [
7
- 0.48145466,
8
- 0.4578275,
9
- 0.40821073
10
- ],
11
- "image_processor_type": "Qwen2VLImageProcessor",
12
- "image_std": [
13
- 0.26862954,
14
- 0.26130258,
15
- 0.27577711
16
- ],
17
- "max_pixels": 12845056,
18
- "merge_size": 2,
19
- "min_pixels": 3136,
20
- "patch_size": 14,
21
- "processor_class": "Qwen2_5_VLProcessor",
22
- "resample": 3,
23
- "rescale_factor": 0.00392156862745098,
24
- "size": {
25
- "longest_edge": 12845056,
26
- "shortest_edge": 3136
27
- },
28
- "temporal_patch_size": 2
29
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/UniWorld-V1/special_tokens_map.json DELETED
@@ -1,31 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "<|im_start|>",
4
- "<|im_end|>",
5
- "<|object_ref_start|>",
6
- "<|object_ref_end|>",
7
- "<|box_start|>",
8
- "<|box_end|>",
9
- "<|quad_start|>",
10
- "<|quad_end|>",
11
- "<|vision_start|>",
12
- "<|vision_end|>",
13
- "<|vision_pad|>",
14
- "<|image_pad|>",
15
- "<|video_pad|>"
16
- ],
17
- "eos_token": {
18
- "content": "<|im_end|>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- },
24
- "pad_token": {
25
- "content": "<|endoftext|>",
26
- "lstrip": false,
27
- "normalized": false,
28
- "rstrip": false,
29
- "single_word": false
30
- }
31
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/UniWorld-V1/task_head_final.pt DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:de8c72b715f1a9f37a36f218b3481e8dbce5fe6f57a7890b660d5da0d611efce
3
- size 146925712
 
 
 
 
models/UniWorld-V1/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:ba0c439f7be467bf47d12a7e6f9adc6116201056fc60c67f431c679b7c16afc8
3
- size 11422064
 
 
 
 
models/UniWorld-V1/tokenizer_config.json DELETED
@@ -1,213 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_prefix_space": false,
4
- "added_tokens_decoder": {
5
- "151643": {
6
- "content": "<|endoftext|>",
7
- "lstrip": false,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "151644": {
14
- "content": "<|im_start|>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": true
20
- },
21
- "151645": {
22
- "content": "<|im_end|>",
23
- "lstrip": false,
24
- "normalized": false,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": true
28
- },
29
- "151646": {
30
- "content": "<|object_ref_start|>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "151647": {
38
- "content": "<|object_ref_end|>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": true
44
- },
45
- "151648": {
46
- "content": "<|box_start|>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": true
52
- },
53
- "151649": {
54
- "content": "<|box_end|>",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": true
60
- },
61
- "151650": {
62
- "content": "<|quad_start|>",
63
- "lstrip": false,
64
- "normalized": false,
65
- "rstrip": false,
66
- "single_word": false,
67
- "special": true
68
- },
69
- "151651": {
70
- "content": "<|quad_end|>",
71
- "lstrip": false,
72
- "normalized": false,
73
- "rstrip": false,
74
- "single_word": false,
75
- "special": true
76
- },
77
- "151652": {
78
- "content": "<|vision_start|>",
79
- "lstrip": false,
80
- "normalized": false,
81
- "rstrip": false,
82
- "single_word": false,
83
- "special": true
84
- },
85
- "151653": {
86
- "content": "<|vision_end|>",
87
- "lstrip": false,
88
- "normalized": false,
89
- "rstrip": false,
90
- "single_word": false,
91
- "special": true
92
- },
93
- "151654": {
94
- "content": "<|vision_pad|>",
95
- "lstrip": false,
96
- "normalized": false,
97
- "rstrip": false,
98
- "single_word": false,
99
- "special": true
100
- },
101
- "151655": {
102
- "content": "<|image_pad|>",
103
- "lstrip": false,
104
- "normalized": false,
105
- "rstrip": false,
106
- "single_word": false,
107
- "special": true
108
- },
109
- "151656": {
110
- "content": "<|video_pad|>",
111
- "lstrip": false,
112
- "normalized": false,
113
- "rstrip": false,
114
- "single_word": false,
115
- "special": true
116
- },
117
- "151657": {
118
- "content": "<tool_call>",
119
- "lstrip": false,
120
- "normalized": false,
121
- "rstrip": false,
122
- "single_word": false,
123
- "special": false
124
- },
125
- "151658": {
126
- "content": "</tool_call>",
127
- "lstrip": false,
128
- "normalized": false,
129
- "rstrip": false,
130
- "single_word": false,
131
- "special": false
132
- },
133
- "151659": {
134
- "content": "<|fim_prefix|>",
135
- "lstrip": false,
136
- "normalized": false,
137
- "rstrip": false,
138
- "single_word": false,
139
- "special": false
140
- },
141
- "151660": {
142
- "content": "<|fim_middle|>",
143
- "lstrip": false,
144
- "normalized": false,
145
- "rstrip": false,
146
- "single_word": false,
147
- "special": false
148
- },
149
- "151661": {
150
- "content": "<|fim_suffix|>",
151
- "lstrip": false,
152
- "normalized": false,
153
- "rstrip": false,
154
- "single_word": false,
155
- "special": false
156
- },
157
- "151662": {
158
- "content": "<|fim_pad|>",
159
- "lstrip": false,
160
- "normalized": false,
161
- "rstrip": false,
162
- "single_word": false,
163
- "special": false
164
- },
165
- "151663": {
166
- "content": "<|repo_name|>",
167
- "lstrip": false,
168
- "normalized": false,
169
- "rstrip": false,
170
- "single_word": false,
171
- "special": false
172
- },
173
- "151664": {
174
- "content": "<|file_sep|>",
175
- "lstrip": false,
176
- "normalized": false,
177
- "rstrip": false,
178
- "single_word": false,
179
- "special": false
180
- }
181
- },
182
- "additional_special_tokens": [
183
- "<|im_start|>",
184
- "<|im_end|>",
185
- "<|object_ref_start|>",
186
- "<|object_ref_end|>",
187
- "<|box_start|>",
188
- "<|box_end|>",
189
- "<|quad_start|>",
190
- "<|quad_end|>",
191
- "<|vision_start|>",
192
- "<|vision_end|>",
193
- "<|vision_pad|>",
194
- "<|image_pad|>",
195
- "<|video_pad|>"
196
- ],
197
- "bos_token": null,
198
- "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
- "clean_up_tokenization_spaces": false,
200
- "eos_token": "<|im_end|>",
201
- "errors": "replace",
202
- "extra_special_tokens": {},
203
- "max_length": null,
204
- "model_max_length": 131072,
205
- "pad_to_multiple_of": null,
206
- "pad_token": "<|endoftext|>",
207
- "pad_token_type_id": 0,
208
- "padding_side": "right",
209
- "processor_class": "Qwen2_5_VLProcessor",
210
- "split_special_tokens": false,
211
- "tokenizer_class": "Qwen2Tokenizer",
212
- "unk_token": null
213
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/UniWorld-V1/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
models/flux/LICENSE.md DELETED
@@ -1,82 +0,0 @@
1
- **FLUX.1 \[dev\] Non-Commercial License**
2
-
3
- Black Forest Labs, Inc. (โ€œ**we**โ€ or โ€œ**our**โ€ or โ€œ**Company**โ€) is pleased to make available the weights, parameters and inference code for the FLUX.1 \[dev\] Model (as defined below) freely available for your non-commercial and non-production use as set forth in this FLUX.1 \[dev\] Non-Commercial License (โ€œ**License**โ€). The โ€œ**FLUX.1 \[dev\] Model**โ€ means the FLUX.1 \[dev\] text-to-image AI model and its elements which includes algorithms, software, checkpoints, parameters, source code (inference code, evaluation code, and if applicable, fine-tuning code) and any other materials associated with the FLUX.1 \[dev\] AI model made available by Company under this License, including if any, the technical documentation, manuals and instructions for the use and operation thereof (collectively, โ€œ**FLUX.1 \[dev\] Model**โ€).
4
-
5
- By downloading, accessing, use, Distributing (as defined below), or creating a Derivative (as defined below) of the FLUX.1 \[dev\] Model, you agree to the terms of this License. If you do not agree to this License, then you do not have any rights to access, use, Distribute or create a Derivative of the FLUX.1 \[dev\] Model and you must immediately cease using the FLUX.1 \[dev\] Model. If you are agreeing to be bound by the terms of this License on behalf of your employer or other entity, you represent and warrant to us that you have full legal authority to bind your employer or such entity to this License. If you do not have the requisite authority, you may not accept the License or access the FLUX.1 \[dev\] Model on behalf of your employer or other entity.
6
-
7
- 1. **Definitions**. Capitalized terms used in this License but not defined herein have the following meanings:
8
-
9
- 1. โ€œ**Derivative**โ€ means any (i) modified version of the FLUX.1 \[dev\] Model (including but not limited to any customized or fine-tuned version thereof), (ii) work based on the FLUX.1 \[dev\] Model, or (iii) any other derivative work thereof. For the avoidance of doubt, Outputs are not considered Derivatives under this License.
10
-
11
- 1. โ€œ**Distribution**โ€ or โ€œ**Distribute**โ€ or โ€œ**Distributing**โ€ means providing or making available, by any means, a copy of the FLUX.1 \[dev\] Models and/or the Derivatives as the case may be.
12
-
13
- 1. โ€œ**Non-Commercial Purpose**โ€ means any of the following uses, but only so far as you do not receive any direct or indirect payment arising from the use of the model or its output: (i) personal use for research, experiment, and testing for the benefit of public knowledge, personal study, private entertainment, hobby projects, or otherwise not directly or indirectly connected to any commercial activities, business operations, or employment responsibilities; (ii) use by commercial or for-profit entities for testing, evaluation, or non-commercial research and development in a non-production environment, (iii) use by any charitable organization for charitable purposes, or for testing or evaluation. For clarity, use for revenue-generating activity or direct interactions with or impacts on end users, or use to train, fine tune or distill other models for commercial use is not a Non-Commercial purpose.
14
-
15
- 1. โ€œ**Outputs**โ€ means any content generated by the operation of the FLUX.1 \[dev\] Models or the Derivatives from a prompt (i.e., text instructions) provided by users. For the avoidance of doubt, Outputs do not include any components of a FLUX.1 \[dev\] Models, such as any fine-tuned versions of the FLUX.1 \[dev\] Models, the weights, or parameters.
16
-
17
- 1. โ€œ**you**โ€ or โ€œ**your**โ€ means the individual or entity entering into this License with Company.
18
-
19
- 1. **License Grant**.
20
-
21
- 1. License. Subject to your compliance with this License, Company grants you a non-exclusive, worldwide, non-transferable, non-sublicensable, revocable, royalty free and limited license to access, use, create Derivatives of, and Distribute the FLUX.1 \[dev\] Models solely for your Non-Commercial Purposes. The foregoing license is personal to you, and you may not assign or sublicense this License or any other rights or obligations under this License without Companyโ€™s prior written consent; any such assignment or sublicense will be void and will automatically and immediately terminate this License. Any restrictions set forth herein in regarding the FLUX.1 \[dev\] Model also applies to any Derivative you create or that are created on your behalf.
22
-
23
- 1. Non-Commercial Use Only. You may only access, use, Distribute, or creative Derivatives of or the FLUX.1 \[dev\] Model or Derivatives for Non-Commercial Purposes. If You want to use a FLUX.1 \[dev\] Model a Derivative for any purpose that is not expressly authorized under this License, such as for a commercial activity, you must request a license from Company, which Company may grant to you in Companyโ€™s sole discretion and which additional use may be subject to a fee, royalty or other revenue share. Please contact Company at the following e-mail address if you want to discuss such a license: [email protected].
24
-
25
- 1. Reserved Rights. The grant of rights expressly set forth in this License are the complete grant of rights to you in the FLUX.1 \[dev\] Model, and no other licenses are granted, whether by waiver, estoppel, implication, equity or otherwise. Company and its licensors reserve all rights not expressly granted by this License.
26
-
27
- 1. Outputs. We claim no ownership rights in and to the Outputs. You are solely responsible for the Outputs you generate and their subsequent uses in accordance with this License. You may use Output for any purpose (including for commercial purposes), except as expressly prohibited herein. You may not use the Output to train, fine-tune or distill a model that is competitive with the FLUX.1 \[dev\] Model.
28
-
29
- 1. **Distribution**. Subject to this License, you may Distribute copies of the FLUX.1 \[dev\] Model and/or Derivatives made by you, under the following conditions:
30
-
31
- 1. you must make available a copy of this License to third-party recipients of the FLUX.1 \[dev\] Models and/or Derivatives you Distribute, and specify that any rights to use the FLUX.1 \[dev\] Models and/or Derivatives shall be directly granted by Company to said third-party recipients pursuant to this License;
32
-
33
- 1. you must make prominently display the following notice alongside the Distribution of the FLUX.1 \[dev\] Model or Derivative (such as via a โ€œNoticeโ€ text file distributed as part of such FLUX.1 \[dev\] Model or Derivative) (the โ€œ**Attribution Notice**โ€):
34
-
35
- โ€œThe FLUX.1 \[dev\] Model is licensed by Black Forest Labs. Inc. under the FLUX.1 \[dev\] Non-Commercial License. Copyright Black Forest Labs. Inc.
36
-
37
- IN NO EVENT SHALL BLACK FOREST LABS, INC. BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH USE OF THIS MODEL.โ€
38
-
39
- 1. in the case of Distribution of Derivatives made by you, you must also include in the Attribution Notice a statement that you have modified the applicable FLUX.1 \[dev\] Model; and
40
-
41
- 1. in the case of Distribution of Derivatives made by you, any terms and conditions you impose on any third-party recipients relating to Derivatives made by or for you shall neither limit such third-party recipientsโ€™ use of the FLUX.1 \[dev\] Model or any Derivatives made by or for Company in accordance with this License nor conflict with any of its terms and conditions.
42
-
43
- 1. In the case of Distribution of Derivatives made by you, you must not misrepresent or imply, through any means, that the Derivatives made by or for you and/or any modified version of the FLUX.1 \[dev\] Model you Distribute under your name and responsibility is an official product of the Company or has been endorsed, approved or validated by the Company, unless you are authorized by Company to do so in writing.
44
-
45
- 1. **Restrictions**. You will not, and will not permit, assist or cause any third party to
46
-
47
- 1. use, modify, copy, reproduce, create Derivatives of, or Distribute the FLUX.1 \[dev\] Model (or any Derivative thereof, or any data produced by the FLUX.1 \[dev\] Model), in whole or in part, for (i) any commercial or production purposes, (ii) military purposes, (iii) purposes of surveillance, including any research or development relating to surveillance, (iv) biometric processing, (v) in any manner that infringes, misappropriates, or otherwise violates any third-party rights, or (vi) in any manner that violates any applicable law and violating any privacy or security laws, rules, regulations, directives, or governmental requirements (including the General Data Privacy Regulation (Regulation (EU) 2016/679), the California Consumer Privacy Act, and any and all laws governing the processing of biometric information), as well as all amendments and successor laws to any of the foregoing;
48
-
49
- 1. alter or remove copyright and other proprietary notices which appear on or in any portion of the FLUX.1 \[dev\] Model;
50
-
51
- 1. utilize any equipment, device, software, or other means to circumvent or remove any security or protection used by Company in connection with the FLUX.1 \[dev\] Model, or to circumvent or remove any usage restrictions, or to enable functionality disabled by FLUX.1 \[dev\] Model; or
52
-
53
- 1. offer or impose any terms on the FLUX.1 \[dev\] Model that alter, restrict, or are inconsistent with the terms of this License.
54
-
55
- 1. violate any applicable U.S. and non-U.S. export control and trade sanctions laws (โ€œ**Export Laws**โ€) in connection with your use or Distribution of any FLUX.1 \[dev\] Model;
56
-
57
- 1. directly or indirectly Distribute, export, or otherwise transfer FLUX.1 \[dev\] Model (a) to any individual, entity, or country prohibited by Export Laws; (b) to anyone on U.S. or non-U.S. government restricted parties lists; or (c) for any purpose prohibited by Export Laws, including nuclear, chemical or biological weapons, or missile technology applications; 3\) use or download FLUX.1 \[dev\] Model if you or they are (a) located in a comprehensively sanctioned jurisdiction, (b) currently listed on any U.S. or non-U.S. restricted parties list, or (c) for any purpose prohibited by Export Laws; and (4) will not disguise your location through IP proxying or other methods.
58
-
59
- 1. **DISCLAIMERS**. THE FLUX.1 \[dev\] MODEL IS PROVIDED โ€œAS ISโ€ AND โ€œWITH ALL FAULTSโ€ WITH NO WARRANTY OF ANY KIND, EXPRESS OR IMPLIED. COMPANY EXPRESSLY DISCLAIMS ALL REPRESENTATIONS AND WARRANTIES, EXPRESS OR IMPLIED, WHETHER BY STATUTE, CUSTOM, USAGE OR OTHERWISE AS TO ANY MATTERS RELATED TO THE FLUX.1 \[dev\] MODEL, INCLUDING BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, TITLE, SATISFACTORY QUALITY, OR NON-INFRINGEMENT. COMPANY MAKES NO WARRANTIES OR REPRESENTATIONS THAT THE FLUX.1 \[dev\] MODEL WILL BE ERROR FREE OR FREE OF VIRUSES OR OTHER HARMFUL COMPONENTS, OR PRODUCE ANY PARTICULAR RESULTS.
60
-
61
- 1. **LIMITATION OF LIABILITY**. TO THE FULLEST EXTENT PERMITTED BY LAW, IN NO EVENT WILL COMPANY BE LIABLE TO YOU OR YOUR EMPLOYEES, AFFILIATES, USERS, OFFICERS OR DIRECTORS (A) UNDER ANY THEORY OF LIABILITY, WHETHER BASED IN CONTRACT, TORT, NEGLIGENCE, STRICT LIABILITY, WARRANTY, OR OTHERWISE UNDER THIS LICENSE, OR (B) FOR ANY INDIRECT, CONSEQUENTIAL, EXEMPLARY, INCIDENTAL, PUNITIVE OR SPECIAL DAMAGES OR LOST PROFITS, EVEN IF COMPANY HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. THE FLUX.1 \[dev\] MODEL, ITS CONSTITUENT COMPONENTS, AND ANY OUTPUT (COLLECTIVELY, โ€œMODEL MATERIALSโ€) ARE NOT DESIGNED OR INTENDED FOR USE IN ANY APPLICATION OR SITUATION WHERE FAILURE OR FAULT OF THE MODEL MATERIALS COULD REASONABLY BE ANTICIPATED TO LEAD TO SERIOUS INJURY OF ANY PERSON, INCLUDING POTENTIAL DISCRIMINATION OR VIOLATION OF AN INDIVIDUALโ€™S PRIVACY RIGHTS, OR TO SEVERE PHYSICAL, PROPERTY, OR ENVIRONMENTAL DAMAGE (EACH, A โ€œ**HIGH-RISK USE**โ€). IF YOU ELECT TO USE ANY OF THE MODEL MATERIALS FOR A HIGH-RISK USE, YOU DO SO AT YOUR OWN RISK. YOU AGREE TO DESIGN AND IMPLEMENT APPROPRIATE DECISION-MAKING AND RISK-MITIGATION PROCEDURES AND POLICIES IN CONNECTION WITH A HIGH-RISK USE SUCH THAT EVEN IF THERE IS A FAILURE OR FAULT IN ANY OF THE MODEL MATERIALS, THE SAFETY OF PERSONS OR PROPERTY AFFECTED BY THE ACTIVITY STAYS AT A LEVEL THAT IS REASONABLE, APPROPRIATE, AND LAWFUL FOR THE FIELD OF THE HIGH-RISK USE.
62
-
63
- 1. **INDEMNIFICATION**
64
-
65
- You will indemnify, defend and hold harmless Company and our subsidiaries and affiliates, and each of our respective shareholders, directors, officers, employees, agents, successors, and assigns (collectively, the โ€œ**Company Parties**โ€) from and against any losses, liabilities, damages, fines, penalties, and expenses (including reasonable attorneysโ€™ fees) incurred by any Company Party in connection with any claim, demand, allegation, lawsuit, proceeding, or investigation (collectively, โ€œ**Claims**โ€) arising out of or related to (a) your access to or use of the FLUX.1 \[dev\] Model (as well as any Output, results or data generated from such access or use), including any High-Risk Use (defined below); (b) your violation of this License; or (c) your violation, misappropriation or infringement of any rights of another (including intellectual property or other proprietary rights and privacy rights). You will promptly notify the Company Parties of any such Claims, and cooperate with Company Parties in defending such Claims. You will also grant the Company Parties sole control of the defense or settlement, at Companyโ€™s sole option, of any Claims. This indemnity is in addition to, and not in lieu of, any other indemnities or remedies set forth in a written agreement between you and Company or the other Company Parties.
66
-
67
- 1. **Termination; Survival**.
68
-
69
- 1. This License will automatically terminate upon any breach by you of the terms of this License.
70
-
71
- 1. We may terminate this License, in whole or in part, at any time upon notice (including electronic) to you.
72
-
73
- 1. If You initiate any legal action or proceedings against Company or any other entity (including a cross-claim or counterclaim in a lawsuit), alleging that the FLUX.1 \[dev\] Model or any Derivative, or any part thereof, infringe upon intellectual property or other rights owned or licensable by you, then any licenses granted to you under this License will immediately terminate as of the date such legal action or claim is filed or initiated.
74
-
75
- 1. Upon termination of this License, you must cease all use, access or Distribution of the FLUX.1 \[dev\] Model and any Derivatives. The following sections survive termination of this License 2(c), 2(d), 4-11.
76
-
77
- 1. **Third Party Materials**. The FLUX.1 \[dev\] Model may contain third-party software or other components (including free and open source software) (all of the foregoing, โ€œ**Third Party Materials**โ€), which are subject to the license terms of the respective third-party licensors. Your dealings or correspondence with third parties and your use of or interaction with any Third Party Materials are solely between you and the third party. Company does not control or endorse, and makes no representations or warranties regarding, any Third Party Materials, and your access to and use of such Third Party Materials are at your own risk.
78
-
79
- 1. **Trademarks**. You have not been granted any trademark license as part of this License and may not use any name or mark associated with Company without the prior written permission of Company, except to the extent necessary to make the reference required in the Attribution Notice as specified above or as is reasonably necessary in describing the FLUX.1 \[dev\] Model and its creators.
80
-
81
- 1. **General**. This License will be governed and construed under the laws of the State of Delaware without regard to conflicts of law provisions. If any provision or part of a provision of this License is unlawful, void or unenforceable, that provision or part of the provision is deemed severed from this License, and will not affect the validity and enforceability of any remaining provisions. The failure of Company to exercise or enforce any right or provision of this License will not operate as a waiver of such right or provision. This License does not confer any third-party beneficiary rights upon any other person or entity. This License, together with the Documentation, contains the entire understanding between you and Company regarding the subject matter of this License, and supersedes all other written or oral agreements and understandings between you and Company regarding such subject matter. No change or addition to any provision of this License will be binding unless it is in writing and signed by an authorized representative of both you and Company.
82
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/model_index.json DELETED
@@ -1,32 +0,0 @@
1
- {
2
- "_class_name": "FluxPipeline",
3
- "_diffusers_version": "0.30.0.dev0",
4
- "scheduler": [
5
- "diffusers",
6
- "FlowMatchEulerDiscreteScheduler"
7
- ],
8
- "text_encoder": [
9
- "transformers",
10
- "CLIPTextModel"
11
- ],
12
- "text_encoder_2": [
13
- "transformers",
14
- "T5EncoderModel"
15
- ],
16
- "tokenizer": [
17
- "transformers",
18
- "CLIPTokenizer"
19
- ],
20
- "tokenizer_2": [
21
- "transformers",
22
- "T5TokenizerFast"
23
- ],
24
- "transformer": [
25
- "diffusers",
26
- "FluxTransformer2DModel"
27
- ],
28
- "vae": [
29
- "diffusers",
30
- "AutoencoderKL"
31
- ]
32
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/put_them_to_flux DELETED
@@ -1 +0,0 @@
1
- https://huggingface.co/black-forest-labs/FLUX.1-dev
 
 
models/flux/scheduler/scheduler_config.json DELETED
@@ -1,11 +0,0 @@
1
- {
2
- "_class_name": "FlowMatchEulerDiscreteScheduler",
3
- "_diffusers_version": "0.30.0.dev0",
4
- "base_image_seq_len": 256,
5
- "base_shift": 0.5,
6
- "max_image_seq_len": 4096,
7
- "max_shift": 1.15,
8
- "num_train_timesteps": 1000,
9
- "shift": 3.0,
10
- "use_dynamic_shifting": true
11
- }
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/text_encoder/config.json DELETED
@@ -1,25 +0,0 @@
1
- {
2
- "_name_or_path": "openai/clip-vit-large-patch14",
3
- "architectures": [
4
- "CLIPTextModel"
5
- ],
6
- "attention_dropout": 0.0,
7
- "bos_token_id": 0,
8
- "dropout": 0.0,
9
- "eos_token_id": 2,
10
- "hidden_act": "quick_gelu",
11
- "hidden_size": 768,
12
- "initializer_factor": 1.0,
13
- "initializer_range": 0.02,
14
- "intermediate_size": 3072,
15
- "layer_norm_eps": 1e-05,
16
- "max_position_embeddings": 77,
17
- "model_type": "clip_text_model",
18
- "num_attention_heads": 12,
19
- "num_hidden_layers": 12,
20
- "pad_token_id": 1,
21
- "projection_dim": 768,
22
- "torch_dtype": "bfloat16",
23
- "transformers_version": "4.43.3",
24
- "vocab_size": 49408
25
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/text_encoder/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:893d67a23f4693ed42cdab4cbad7fe3e727cf59609c40da28a46b5470f9ed082
3
- size 246144352
 
 
 
 
models/flux/text_encoder_2/config.json DELETED
@@ -1,22 +0,0 @@
1
- {
2
- "d_ff": 10240,
3
- "d_kv": 64,
4
- "d_model": 4096,
5
- "decoder_start_token_id": 0,
6
- "dropout_rate": 0.1,
7
- "eos_token_id": 1,
8
- "dense_act_fn": "gelu_pytorch_tanh",
9
- "initializer_factor": 1.0,
10
- "is_encoder_decoder": true,
11
- "is_gated_act": true,
12
- "layer_norm_epsilon": 1e-06,
13
- "model_type": "t5",
14
- "num_decoder_layers": 24,
15
- "num_heads": 64,
16
- "num_layers": 24,
17
- "output_past": true,
18
- "pad_token_id": 0,
19
- "relative_attention_num_buckets": 32,
20
- "tie_word_embeddings": false,
21
- "vocab_size": 69328
22
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/text_encoder_2/model.safetensors.index.json DELETED
@@ -1,226 +0,0 @@
1
- {
2
- "metadata": {
3
- "total_size": 10397325856
4
- },
5
- "weight_map": {
6
- "encoder.block.0.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
7
- "encoder.block.0.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
8
- "encoder.block.0.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
9
- "encoder.block.0.layer.0.SelfAttention.relative_attention_bias.weight": "t5xxl-unchained-f16.safetensors",
10
- "encoder.block.0.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
11
- "encoder.block.0.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
12
- "encoder.block.0.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
13
- "encoder.block.0.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
14
- "encoder.block.0.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
15
- "encoder.block.0.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
16
- "encoder.block.1.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
17
- "encoder.block.1.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
18
- "encoder.block.1.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
19
- "encoder.block.1.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
20
- "encoder.block.1.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
21
- "encoder.block.1.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
22
- "encoder.block.1.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
23
- "encoder.block.1.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
24
- "encoder.block.1.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
25
- "encoder.block.10.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
26
- "encoder.block.10.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
27
- "encoder.block.10.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
28
- "encoder.block.10.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
29
- "encoder.block.10.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
30
- "encoder.block.10.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
31
- "encoder.block.10.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
32
- "encoder.block.10.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
33
- "encoder.block.10.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
34
- "encoder.block.11.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
35
- "encoder.block.11.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
36
- "encoder.block.11.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
37
- "encoder.block.11.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
38
- "encoder.block.11.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
39
- "encoder.block.11.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
40
- "encoder.block.11.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
41
- "encoder.block.11.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
42
- "encoder.block.11.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
43
- "encoder.block.12.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
44
- "encoder.block.12.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
45
- "encoder.block.12.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
46
- "encoder.block.12.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
47
- "encoder.block.12.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
48
- "encoder.block.12.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
49
- "encoder.block.12.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
50
- "encoder.block.12.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
51
- "encoder.block.12.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
52
- "encoder.block.13.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
53
- "encoder.block.13.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
54
- "encoder.block.13.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
55
- "encoder.block.13.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
56
- "encoder.block.13.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
57
- "encoder.block.13.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
58
- "encoder.block.13.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
59
- "encoder.block.13.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
60
- "encoder.block.13.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
61
- "encoder.block.14.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
62
- "encoder.block.14.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
63
- "encoder.block.14.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
64
- "encoder.block.14.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
65
- "encoder.block.14.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
66
- "encoder.block.14.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
67
- "encoder.block.14.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
68
- "encoder.block.14.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
69
- "encoder.block.14.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
70
- "encoder.block.15.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
71
- "encoder.block.15.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
72
- "encoder.block.15.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
73
- "encoder.block.15.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
74
- "encoder.block.15.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
75
- "encoder.block.15.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
76
- "encoder.block.15.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
77
- "encoder.block.15.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
78
- "encoder.block.15.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
79
- "encoder.block.16.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
80
- "encoder.block.16.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
81
- "encoder.block.16.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
82
- "encoder.block.16.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
83
- "encoder.block.16.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
84
- "encoder.block.16.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
85
- "encoder.block.16.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
86
- "encoder.block.16.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
87
- "encoder.block.16.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
88
- "encoder.block.17.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
89
- "encoder.block.17.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
90
- "encoder.block.17.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
91
- "encoder.block.17.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
92
- "encoder.block.17.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
93
- "encoder.block.17.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
94
- "encoder.block.17.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
95
- "encoder.block.17.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
96
- "encoder.block.17.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
97
- "encoder.block.18.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
98
- "encoder.block.18.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
99
- "encoder.block.18.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
100
- "encoder.block.18.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
101
- "encoder.block.18.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
102
- "encoder.block.18.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
103
- "encoder.block.18.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
104
- "encoder.block.18.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
105
- "encoder.block.18.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
106
- "encoder.block.19.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
107
- "encoder.block.19.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
108
- "encoder.block.19.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
109
- "encoder.block.19.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
110
- "encoder.block.19.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
111
- "encoder.block.19.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
112
- "encoder.block.19.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
113
- "encoder.block.19.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
114
- "encoder.block.19.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
115
- "encoder.block.2.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
116
- "encoder.block.2.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
117
- "encoder.block.2.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
118
- "encoder.block.2.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
119
- "encoder.block.2.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
120
- "encoder.block.2.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
121
- "encoder.block.2.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
122
- "encoder.block.2.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
123
- "encoder.block.2.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
124
- "encoder.block.20.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
125
- "encoder.block.20.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
126
- "encoder.block.20.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
127
- "encoder.block.20.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
128
- "encoder.block.20.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
129
- "encoder.block.20.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
130
- "encoder.block.20.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
131
- "encoder.block.20.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
132
- "encoder.block.20.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
133
- "encoder.block.21.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
134
- "encoder.block.21.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
135
- "encoder.block.21.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
136
- "encoder.block.21.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
137
- "encoder.block.21.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
138
- "encoder.block.21.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
139
- "encoder.block.21.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
140
- "encoder.block.21.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
141
- "encoder.block.21.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
142
- "encoder.block.22.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
143
- "encoder.block.22.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
144
- "encoder.block.22.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
145
- "encoder.block.22.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
146
- "encoder.block.22.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
147
- "encoder.block.22.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
148
- "encoder.block.22.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
149
- "encoder.block.22.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
150
- "encoder.block.22.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
151
- "encoder.block.23.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
152
- "encoder.block.23.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
153
- "encoder.block.23.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
154
- "encoder.block.23.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
155
- "encoder.block.23.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
156
- "encoder.block.23.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
157
- "encoder.block.23.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
158
- "encoder.block.23.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
159
- "encoder.block.23.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
160
- "encoder.block.3.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
161
- "encoder.block.3.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
162
- "encoder.block.3.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
163
- "encoder.block.3.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
164
- "encoder.block.3.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
165
- "encoder.block.3.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
166
- "encoder.block.3.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
167
- "encoder.block.3.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
168
- "encoder.block.3.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
169
- "encoder.block.4.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
170
- "encoder.block.4.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
171
- "encoder.block.4.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
172
- "encoder.block.4.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
173
- "encoder.block.4.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
174
- "encoder.block.4.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
175
- "encoder.block.4.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
176
- "encoder.block.4.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
177
- "encoder.block.4.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
178
- "encoder.block.5.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
179
- "encoder.block.5.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
180
- "encoder.block.5.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
181
- "encoder.block.5.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
182
- "encoder.block.5.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
183
- "encoder.block.5.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
184
- "encoder.block.5.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
185
- "encoder.block.5.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
186
- "encoder.block.5.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
187
- "encoder.block.6.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
188
- "encoder.block.6.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
189
- "encoder.block.6.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
190
- "encoder.block.6.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
191
- "encoder.block.6.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
192
- "encoder.block.6.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
193
- "encoder.block.6.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
194
- "encoder.block.6.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
195
- "encoder.block.6.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
196
- "encoder.block.7.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
197
- "encoder.block.7.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
198
- "encoder.block.7.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
199
- "encoder.block.7.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
200
- "encoder.block.7.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
201
- "encoder.block.7.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
202
- "encoder.block.7.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
203
- "encoder.block.7.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
204
- "encoder.block.7.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
205
- "encoder.block.8.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
206
- "encoder.block.8.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
207
- "encoder.block.8.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
208
- "encoder.block.8.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
209
- "encoder.block.8.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
210
- "encoder.block.8.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
211
- "encoder.block.8.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
212
- "encoder.block.8.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
213
- "encoder.block.8.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
214
- "encoder.block.9.layer.0.SelfAttention.k.weight": "t5xxl-unchained-f16.safetensors",
215
- "encoder.block.9.layer.0.SelfAttention.o.weight": "t5xxl-unchained-f16.safetensors",
216
- "encoder.block.9.layer.0.SelfAttention.q.weight": "t5xxl-unchained-f16.safetensors",
217
- "encoder.block.9.layer.0.SelfAttention.v.weight": "t5xxl-unchained-f16.safetensors",
218
- "encoder.block.9.layer.0.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
219
- "encoder.block.9.layer.1.DenseReluDense.wi_0.weight": "t5xxl-unchained-f16.safetensors",
220
- "encoder.block.9.layer.1.DenseReluDense.wi_1.weight": "t5xxl-unchained-f16.safetensors",
221
- "encoder.block.9.layer.1.DenseReluDense.wo.weight": "t5xxl-unchained-f16.safetensors",
222
- "encoder.block.9.layer.1.layer_norm.weight": "t5xxl-unchained-f16.safetensors",
223
- "encoder.final_layer_norm.weight": "t5xxl-unchained-f16.safetensors",
224
- "shared.weight": "t5xxl-unchained-f16.safetensors"
225
- }
226
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/text_encoder_2/t5xxl-unchained-f16.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:53a6cee6c313b938f21c6fa983f444326e9988b016385d1623aea7b166b9cebe
3
- size 10397325856
 
 
 
 
models/flux/tokenizer/merges.txt DELETED
The diff for this file is too large to render. See raw diff
 
models/flux/tokenizer/special_tokens_map.json DELETED
@@ -1,30 +0,0 @@
1
- {
2
- "bos_token": {
3
- "content": "<|startoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": false,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<|endoftext|>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
- "unk_token": {
24
- "content": "<|endoftext|>",
25
- "lstrip": false,
26
- "normalized": false,
27
- "rstrip": false,
28
- "single_word": false
29
- }
30
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/tokenizer/tokenizer_config.json DELETED
@@ -1,30 +0,0 @@
1
- {
2
- "add_prefix_space": false,
3
- "added_tokens_decoder": {
4
- "49406": {
5
- "content": "<|startoftext|>",
6
- "lstrip": false,
7
- "normalized": true,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": true
11
- },
12
- "49407": {
13
- "content": "<|endoftext|>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- }
20
- },
21
- "bos_token": "<|startoftext|>",
22
- "clean_up_tokenization_spaces": true,
23
- "do_lower_case": true,
24
- "eos_token": "<|endoftext|>",
25
- "errors": "replace",
26
- "model_max_length": 77,
27
- "pad_token": "<|endoftext|>",
28
- "tokenizer_class": "CLIPTokenizer",
29
- "unk_token": "<|endoftext|>"
30
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/tokenizer/vocab.json DELETED
The diff for this file is too large to render. See raw diff
 
models/flux/tokenizer_2/special_tokens_map.json DELETED
@@ -1,125 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "<extra_id_0>",
4
- "<extra_id_1>",
5
- "<extra_id_2>",
6
- "<extra_id_3>",
7
- "<extra_id_4>",
8
- "<extra_id_5>",
9
- "<extra_id_6>",
10
- "<extra_id_7>",
11
- "<extra_id_8>",
12
- "<extra_id_9>",
13
- "<extra_id_10>",
14
- "<extra_id_11>",
15
- "<extra_id_12>",
16
- "<extra_id_13>",
17
- "<extra_id_14>",
18
- "<extra_id_15>",
19
- "<extra_id_16>",
20
- "<extra_id_17>",
21
- "<extra_id_18>",
22
- "<extra_id_19>",
23
- "<extra_id_20>",
24
- "<extra_id_21>",
25
- "<extra_id_22>",
26
- "<extra_id_23>",
27
- "<extra_id_24>",
28
- "<extra_id_25>",
29
- "<extra_id_26>",
30
- "<extra_id_27>",
31
- "<extra_id_28>",
32
- "<extra_id_29>",
33
- "<extra_id_30>",
34
- "<extra_id_31>",
35
- "<extra_id_32>",
36
- "<extra_id_33>",
37
- "<extra_id_34>",
38
- "<extra_id_35>",
39
- "<extra_id_36>",
40
- "<extra_id_37>",
41
- "<extra_id_38>",
42
- "<extra_id_39>",
43
- "<extra_id_40>",
44
- "<extra_id_41>",
45
- "<extra_id_42>",
46
- "<extra_id_43>",
47
- "<extra_id_44>",
48
- "<extra_id_45>",
49
- "<extra_id_46>",
50
- "<extra_id_47>",
51
- "<extra_id_48>",
52
- "<extra_id_49>",
53
- "<extra_id_50>",
54
- "<extra_id_51>",
55
- "<extra_id_52>",
56
- "<extra_id_53>",
57
- "<extra_id_54>",
58
- "<extra_id_55>",
59
- "<extra_id_56>",
60
- "<extra_id_57>",
61
- "<extra_id_58>",
62
- "<extra_id_59>",
63
- "<extra_id_60>",
64
- "<extra_id_61>",
65
- "<extra_id_62>",
66
- "<extra_id_63>",
67
- "<extra_id_64>",
68
- "<extra_id_65>",
69
- "<extra_id_66>",
70
- "<extra_id_67>",
71
- "<extra_id_68>",
72
- "<extra_id_69>",
73
- "<extra_id_70>",
74
- "<extra_id_71>",
75
- "<extra_id_72>",
76
- "<extra_id_73>",
77
- "<extra_id_74>",
78
- "<extra_id_75>",
79
- "<extra_id_76>",
80
- "<extra_id_77>",
81
- "<extra_id_78>",
82
- "<extra_id_79>",
83
- "<extra_id_80>",
84
- "<extra_id_81>",
85
- "<extra_id_82>",
86
- "<extra_id_83>",
87
- "<extra_id_84>",
88
- "<extra_id_85>",
89
- "<extra_id_86>",
90
- "<extra_id_87>",
91
- "<extra_id_88>",
92
- "<extra_id_89>",
93
- "<extra_id_90>",
94
- "<extra_id_91>",
95
- "<extra_id_92>",
96
- "<extra_id_93>",
97
- "<extra_id_94>",
98
- "<extra_id_95>",
99
- "<extra_id_96>",
100
- "<extra_id_97>",
101
- "<extra_id_98>",
102
- "<extra_id_99>"
103
- ],
104
- "eos_token": {
105
- "content": "</s>",
106
- "lstrip": false,
107
- "normalized": false,
108
- "rstrip": false,
109
- "single_word": false
110
- },
111
- "pad_token": {
112
- "content": "<pad>",
113
- "lstrip": false,
114
- "normalized": false,
115
- "rstrip": false,
116
- "single_word": false
117
- },
118
- "unk_token": {
119
- "content": "<unk>",
120
- "lstrip": false,
121
- "normalized": false,
122
- "rstrip": false,
123
- "single_word": false
124
- }
125
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/tokenizer_2/spiece.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
- size 791656
 
 
 
 
models/flux/tokenizer_2/tokenizer.json DELETED
The diff for this file is too large to render. See raw diff
 
models/flux/tokenizer_2/tokenizer1.json DELETED
The diff for this file is too large to render. See raw diff
 
models/flux/tokenizer_2/tokenizer_config.json DELETED
@@ -1,940 +0,0 @@
1
- {
2
- "add_prefix_space": true,
3
- "added_tokens_decoder": {
4
- "0": {
5
- "content": "<pad>",
6
- "lstrip": false,
7
- "normalized": false,
8
- "rstrip": false,
9
- "single_word": false,
10
- "special": true
11
- },
12
- "1": {
13
- "content": "</s>",
14
- "lstrip": false,
15
- "normalized": false,
16
- "rstrip": false,
17
- "single_word": false,
18
- "special": true
19
- },
20
- "2": {
21
- "content": "<unk>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false,
26
- "special": true
27
- },
28
- "32000": {
29
- "content": "<extra_id_99>",
30
- "lstrip": false,
31
- "normalized": false,
32
- "rstrip": false,
33
- "single_word": false,
34
- "special": true
35
- },
36
- "32001": {
37
- "content": "<extra_id_98>",
38
- "lstrip": false,
39
- "normalized": false,
40
- "rstrip": false,
41
- "single_word": false,
42
- "special": true
43
- },
44
- "32002": {
45
- "content": "<extra_id_97>",
46
- "lstrip": false,
47
- "normalized": false,
48
- "rstrip": false,
49
- "single_word": false,
50
- "special": true
51
- },
52
- "32003": {
53
- "content": "<extra_id_96>",
54
- "lstrip": false,
55
- "normalized": false,
56
- "rstrip": false,
57
- "single_word": false,
58
- "special": true
59
- },
60
- "32004": {
61
- "content": "<extra_id_95>",
62
- "lstrip": false,
63
- "normalized": false,
64
- "rstrip": false,
65
- "single_word": false,
66
- "special": true
67
- },
68
- "32005": {
69
- "content": "<extra_id_94>",
70
- "lstrip": false,
71
- "normalized": false,
72
- "rstrip": false,
73
- "single_word": false,
74
- "special": true
75
- },
76
- "32006": {
77
- "content": "<extra_id_93>",
78
- "lstrip": false,
79
- "normalized": false,
80
- "rstrip": false,
81
- "single_word": false,
82
- "special": true
83
- },
84
- "32007": {
85
- "content": "<extra_id_92>",
86
- "lstrip": false,
87
- "normalized": false,
88
- "rstrip": false,
89
- "single_word": false,
90
- "special": true
91
- },
92
- "32008": {
93
- "content": "<extra_id_91>",
94
- "lstrip": false,
95
- "normalized": false,
96
- "rstrip": false,
97
- "single_word": false,
98
- "special": true
99
- },
100
- "32009": {
101
- "content": "<extra_id_90>",
102
- "lstrip": false,
103
- "normalized": false,
104
- "rstrip": false,
105
- "single_word": false,
106
- "special": true
107
- },
108
- "32010": {
109
- "content": "<extra_id_89>",
110
- "lstrip": false,
111
- "normalized": false,
112
- "rstrip": false,
113
- "single_word": false,
114
- "special": true
115
- },
116
- "32011": {
117
- "content": "<extra_id_88>",
118
- "lstrip": false,
119
- "normalized": false,
120
- "rstrip": false,
121
- "single_word": false,
122
- "special": true
123
- },
124
- "32012": {
125
- "content": "<extra_id_87>",
126
- "lstrip": false,
127
- "normalized": false,
128
- "rstrip": false,
129
- "single_word": false,
130
- "special": true
131
- },
132
- "32013": {
133
- "content": "<extra_id_86>",
134
- "lstrip": false,
135
- "normalized": false,
136
- "rstrip": false,
137
- "single_word": false,
138
- "special": true
139
- },
140
- "32014": {
141
- "content": "<extra_id_85>",
142
- "lstrip": false,
143
- "normalized": false,
144
- "rstrip": false,
145
- "single_word": false,
146
- "special": true
147
- },
148
- "32015": {
149
- "content": "<extra_id_84>",
150
- "lstrip": false,
151
- "normalized": false,
152
- "rstrip": false,
153
- "single_word": false,
154
- "special": true
155
- },
156
- "32016": {
157
- "content": "<extra_id_83>",
158
- "lstrip": false,
159
- "normalized": false,
160
- "rstrip": false,
161
- "single_word": false,
162
- "special": true
163
- },
164
- "32017": {
165
- "content": "<extra_id_82>",
166
- "lstrip": false,
167
- "normalized": false,
168
- "rstrip": false,
169
- "single_word": false,
170
- "special": true
171
- },
172
- "32018": {
173
- "content": "<extra_id_81>",
174
- "lstrip": false,
175
- "normalized": false,
176
- "rstrip": false,
177
- "single_word": false,
178
- "special": true
179
- },
180
- "32019": {
181
- "content": "<extra_id_80>",
182
- "lstrip": false,
183
- "normalized": false,
184
- "rstrip": false,
185
- "single_word": false,
186
- "special": true
187
- },
188
- "32020": {
189
- "content": "<extra_id_79>",
190
- "lstrip": false,
191
- "normalized": false,
192
- "rstrip": false,
193
- "single_word": false,
194
- "special": true
195
- },
196
- "32021": {
197
- "content": "<extra_id_78>",
198
- "lstrip": false,
199
- "normalized": false,
200
- "rstrip": false,
201
- "single_word": false,
202
- "special": true
203
- },
204
- "32022": {
205
- "content": "<extra_id_77>",
206
- "lstrip": false,
207
- "normalized": false,
208
- "rstrip": false,
209
- "single_word": false,
210
- "special": true
211
- },
212
- "32023": {
213
- "content": "<extra_id_76>",
214
- "lstrip": false,
215
- "normalized": false,
216
- "rstrip": false,
217
- "single_word": false,
218
- "special": true
219
- },
220
- "32024": {
221
- "content": "<extra_id_75>",
222
- "lstrip": false,
223
- "normalized": false,
224
- "rstrip": false,
225
- "single_word": false,
226
- "special": true
227
- },
228
- "32025": {
229
- "content": "<extra_id_74>",
230
- "lstrip": false,
231
- "normalized": false,
232
- "rstrip": false,
233
- "single_word": false,
234
- "special": true
235
- },
236
- "32026": {
237
- "content": "<extra_id_73>",
238
- "lstrip": false,
239
- "normalized": false,
240
- "rstrip": false,
241
- "single_word": false,
242
- "special": true
243
- },
244
- "32027": {
245
- "content": "<extra_id_72>",
246
- "lstrip": false,
247
- "normalized": false,
248
- "rstrip": false,
249
- "single_word": false,
250
- "special": true
251
- },
252
- "32028": {
253
- "content": "<extra_id_71>",
254
- "lstrip": false,
255
- "normalized": false,
256
- "rstrip": false,
257
- "single_word": false,
258
- "special": true
259
- },
260
- "32029": {
261
- "content": "<extra_id_70>",
262
- "lstrip": false,
263
- "normalized": false,
264
- "rstrip": false,
265
- "single_word": false,
266
- "special": true
267
- },
268
- "32030": {
269
- "content": "<extra_id_69>",
270
- "lstrip": false,
271
- "normalized": false,
272
- "rstrip": false,
273
- "single_word": false,
274
- "special": true
275
- },
276
- "32031": {
277
- "content": "<extra_id_68>",
278
- "lstrip": false,
279
- "normalized": false,
280
- "rstrip": false,
281
- "single_word": false,
282
- "special": true
283
- },
284
- "32032": {
285
- "content": "<extra_id_67>",
286
- "lstrip": false,
287
- "normalized": false,
288
- "rstrip": false,
289
- "single_word": false,
290
- "special": true
291
- },
292
- "32033": {
293
- "content": "<extra_id_66>",
294
- "lstrip": false,
295
- "normalized": false,
296
- "rstrip": false,
297
- "single_word": false,
298
- "special": true
299
- },
300
- "32034": {
301
- "content": "<extra_id_65>",
302
- "lstrip": false,
303
- "normalized": false,
304
- "rstrip": false,
305
- "single_word": false,
306
- "special": true
307
- },
308
- "32035": {
309
- "content": "<extra_id_64>",
310
- "lstrip": false,
311
- "normalized": false,
312
- "rstrip": false,
313
- "single_word": false,
314
- "special": true
315
- },
316
- "32036": {
317
- "content": "<extra_id_63>",
318
- "lstrip": false,
319
- "normalized": false,
320
- "rstrip": false,
321
- "single_word": false,
322
- "special": true
323
- },
324
- "32037": {
325
- "content": "<extra_id_62>",
326
- "lstrip": false,
327
- "normalized": false,
328
- "rstrip": false,
329
- "single_word": false,
330
- "special": true
331
- },
332
- "32038": {
333
- "content": "<extra_id_61>",
334
- "lstrip": false,
335
- "normalized": false,
336
- "rstrip": false,
337
- "single_word": false,
338
- "special": true
339
- },
340
- "32039": {
341
- "content": "<extra_id_60>",
342
- "lstrip": false,
343
- "normalized": false,
344
- "rstrip": false,
345
- "single_word": false,
346
- "special": true
347
- },
348
- "32040": {
349
- "content": "<extra_id_59>",
350
- "lstrip": false,
351
- "normalized": false,
352
- "rstrip": false,
353
- "single_word": false,
354
- "special": true
355
- },
356
- "32041": {
357
- "content": "<extra_id_58>",
358
- "lstrip": false,
359
- "normalized": false,
360
- "rstrip": false,
361
- "single_word": false,
362
- "special": true
363
- },
364
- "32042": {
365
- "content": "<extra_id_57>",
366
- "lstrip": false,
367
- "normalized": false,
368
- "rstrip": false,
369
- "single_word": false,
370
- "special": true
371
- },
372
- "32043": {
373
- "content": "<extra_id_56>",
374
- "lstrip": false,
375
- "normalized": false,
376
- "rstrip": false,
377
- "single_word": false,
378
- "special": true
379
- },
380
- "32044": {
381
- "content": "<extra_id_55>",
382
- "lstrip": false,
383
- "normalized": false,
384
- "rstrip": false,
385
- "single_word": false,
386
- "special": true
387
- },
388
- "32045": {
389
- "content": "<extra_id_54>",
390
- "lstrip": false,
391
- "normalized": false,
392
- "rstrip": false,
393
- "single_word": false,
394
- "special": true
395
- },
396
- "32046": {
397
- "content": "<extra_id_53>",
398
- "lstrip": false,
399
- "normalized": false,
400
- "rstrip": false,
401
- "single_word": false,
402
- "special": true
403
- },
404
- "32047": {
405
- "content": "<extra_id_52>",
406
- "lstrip": false,
407
- "normalized": false,
408
- "rstrip": false,
409
- "single_word": false,
410
- "special": true
411
- },
412
- "32048": {
413
- "content": "<extra_id_51>",
414
- "lstrip": false,
415
- "normalized": false,
416
- "rstrip": false,
417
- "single_word": false,
418
- "special": true
419
- },
420
- "32049": {
421
- "content": "<extra_id_50>",
422
- "lstrip": false,
423
- "normalized": false,
424
- "rstrip": false,
425
- "single_word": false,
426
- "special": true
427
- },
428
- "32050": {
429
- "content": "<extra_id_49>",
430
- "lstrip": false,
431
- "normalized": false,
432
- "rstrip": false,
433
- "single_word": false,
434
- "special": true
435
- },
436
- "32051": {
437
- "content": "<extra_id_48>",
438
- "lstrip": false,
439
- "normalized": false,
440
- "rstrip": false,
441
- "single_word": false,
442
- "special": true
443
- },
444
- "32052": {
445
- "content": "<extra_id_47>",
446
- "lstrip": false,
447
- "normalized": false,
448
- "rstrip": false,
449
- "single_word": false,
450
- "special": true
451
- },
452
- "32053": {
453
- "content": "<extra_id_46>",
454
- "lstrip": false,
455
- "normalized": false,
456
- "rstrip": false,
457
- "single_word": false,
458
- "special": true
459
- },
460
- "32054": {
461
- "content": "<extra_id_45>",
462
- "lstrip": false,
463
- "normalized": false,
464
- "rstrip": false,
465
- "single_word": false,
466
- "special": true
467
- },
468
- "32055": {
469
- "content": "<extra_id_44>",
470
- "lstrip": false,
471
- "normalized": false,
472
- "rstrip": false,
473
- "single_word": false,
474
- "special": true
475
- },
476
- "32056": {
477
- "content": "<extra_id_43>",
478
- "lstrip": false,
479
- "normalized": false,
480
- "rstrip": false,
481
- "single_word": false,
482
- "special": true
483
- },
484
- "32057": {
485
- "content": "<extra_id_42>",
486
- "lstrip": false,
487
- "normalized": false,
488
- "rstrip": false,
489
- "single_word": false,
490
- "special": true
491
- },
492
- "32058": {
493
- "content": "<extra_id_41>",
494
- "lstrip": false,
495
- "normalized": false,
496
- "rstrip": false,
497
- "single_word": false,
498
- "special": true
499
- },
500
- "32059": {
501
- "content": "<extra_id_40>",
502
- "lstrip": false,
503
- "normalized": false,
504
- "rstrip": false,
505
- "single_word": false,
506
- "special": true
507
- },
508
- "32060": {
509
- "content": "<extra_id_39>",
510
- "lstrip": false,
511
- "normalized": false,
512
- "rstrip": false,
513
- "single_word": false,
514
- "special": true
515
- },
516
- "32061": {
517
- "content": "<extra_id_38>",
518
- "lstrip": false,
519
- "normalized": false,
520
- "rstrip": false,
521
- "single_word": false,
522
- "special": true
523
- },
524
- "32062": {
525
- "content": "<extra_id_37>",
526
- "lstrip": false,
527
- "normalized": false,
528
- "rstrip": false,
529
- "single_word": false,
530
- "special": true
531
- },
532
- "32063": {
533
- "content": "<extra_id_36>",
534
- "lstrip": false,
535
- "normalized": false,
536
- "rstrip": false,
537
- "single_word": false,
538
- "special": true
539
- },
540
- "32064": {
541
- "content": "<extra_id_35>",
542
- "lstrip": false,
543
- "normalized": false,
544
- "rstrip": false,
545
- "single_word": false,
546
- "special": true
547
- },
548
- "32065": {
549
- "content": "<extra_id_34>",
550
- "lstrip": false,
551
- "normalized": false,
552
- "rstrip": false,
553
- "single_word": false,
554
- "special": true
555
- },
556
- "32066": {
557
- "content": "<extra_id_33>",
558
- "lstrip": false,
559
- "normalized": false,
560
- "rstrip": false,
561
- "single_word": false,
562
- "special": true
563
- },
564
- "32067": {
565
- "content": "<extra_id_32>",
566
- "lstrip": false,
567
- "normalized": false,
568
- "rstrip": false,
569
- "single_word": false,
570
- "special": true
571
- },
572
- "32068": {
573
- "content": "<extra_id_31>",
574
- "lstrip": false,
575
- "normalized": false,
576
- "rstrip": false,
577
- "single_word": false,
578
- "special": true
579
- },
580
- "32069": {
581
- "content": "<extra_id_30>",
582
- "lstrip": false,
583
- "normalized": false,
584
- "rstrip": false,
585
- "single_word": false,
586
- "special": true
587
- },
588
- "32070": {
589
- "content": "<extra_id_29>",
590
- "lstrip": false,
591
- "normalized": false,
592
- "rstrip": false,
593
- "single_word": false,
594
- "special": true
595
- },
596
- "32071": {
597
- "content": "<extra_id_28>",
598
- "lstrip": false,
599
- "normalized": false,
600
- "rstrip": false,
601
- "single_word": false,
602
- "special": true
603
- },
604
- "32072": {
605
- "content": "<extra_id_27>",
606
- "lstrip": false,
607
- "normalized": false,
608
- "rstrip": false,
609
- "single_word": false,
610
- "special": true
611
- },
612
- "32073": {
613
- "content": "<extra_id_26>",
614
- "lstrip": false,
615
- "normalized": false,
616
- "rstrip": false,
617
- "single_word": false,
618
- "special": true
619
- },
620
- "32074": {
621
- "content": "<extra_id_25>",
622
- "lstrip": false,
623
- "normalized": false,
624
- "rstrip": false,
625
- "single_word": false,
626
- "special": true
627
- },
628
- "32075": {
629
- "content": "<extra_id_24>",
630
- "lstrip": false,
631
- "normalized": false,
632
- "rstrip": false,
633
- "single_word": false,
634
- "special": true
635
- },
636
- "32076": {
637
- "content": "<extra_id_23>",
638
- "lstrip": false,
639
- "normalized": false,
640
- "rstrip": false,
641
- "single_word": false,
642
- "special": true
643
- },
644
- "32077": {
645
- "content": "<extra_id_22>",
646
- "lstrip": false,
647
- "normalized": false,
648
- "rstrip": false,
649
- "single_word": false,
650
- "special": true
651
- },
652
- "32078": {
653
- "content": "<extra_id_21>",
654
- "lstrip": false,
655
- "normalized": false,
656
- "rstrip": false,
657
- "single_word": false,
658
- "special": true
659
- },
660
- "32079": {
661
- "content": "<extra_id_20>",
662
- "lstrip": false,
663
- "normalized": false,
664
- "rstrip": false,
665
- "single_word": false,
666
- "special": true
667
- },
668
- "32080": {
669
- "content": "<extra_id_19>",
670
- "lstrip": false,
671
- "normalized": false,
672
- "rstrip": false,
673
- "single_word": false,
674
- "special": true
675
- },
676
- "32081": {
677
- "content": "<extra_id_18>",
678
- "lstrip": false,
679
- "normalized": false,
680
- "rstrip": false,
681
- "single_word": false,
682
- "special": true
683
- },
684
- "32082": {
685
- "content": "<extra_id_17>",
686
- "lstrip": false,
687
- "normalized": false,
688
- "rstrip": false,
689
- "single_word": false,
690
- "special": true
691
- },
692
- "32083": {
693
- "content": "<extra_id_16>",
694
- "lstrip": false,
695
- "normalized": false,
696
- "rstrip": false,
697
- "single_word": false,
698
- "special": true
699
- },
700
- "32084": {
701
- "content": "<extra_id_15>",
702
- "lstrip": false,
703
- "normalized": false,
704
- "rstrip": false,
705
- "single_word": false,
706
- "special": true
707
- },
708
- "32085": {
709
- "content": "<extra_id_14>",
710
- "lstrip": false,
711
- "normalized": false,
712
- "rstrip": false,
713
- "single_word": false,
714
- "special": true
715
- },
716
- "32086": {
717
- "content": "<extra_id_13>",
718
- "lstrip": false,
719
- "normalized": false,
720
- "rstrip": false,
721
- "single_word": false,
722
- "special": true
723
- },
724
- "32087": {
725
- "content": "<extra_id_12>",
726
- "lstrip": false,
727
- "normalized": false,
728
- "rstrip": false,
729
- "single_word": false,
730
- "special": true
731
- },
732
- "32088": {
733
- "content": "<extra_id_11>",
734
- "lstrip": false,
735
- "normalized": false,
736
- "rstrip": false,
737
- "single_word": false,
738
- "special": true
739
- },
740
- "32089": {
741
- "content": "<extra_id_10>",
742
- "lstrip": false,
743
- "normalized": false,
744
- "rstrip": false,
745
- "single_word": false,
746
- "special": true
747
- },
748
- "32090": {
749
- "content": "<extra_id_9>",
750
- "lstrip": false,
751
- "normalized": false,
752
- "rstrip": false,
753
- "single_word": false,
754
- "special": true
755
- },
756
- "32091": {
757
- "content": "<extra_id_8>",
758
- "lstrip": false,
759
- "normalized": false,
760
- "rstrip": false,
761
- "single_word": false,
762
- "special": true
763
- },
764
- "32092": {
765
- "content": "<extra_id_7>",
766
- "lstrip": false,
767
- "normalized": false,
768
- "rstrip": false,
769
- "single_word": false,
770
- "special": true
771
- },
772
- "32093": {
773
- "content": "<extra_id_6>",
774
- "lstrip": false,
775
- "normalized": false,
776
- "rstrip": false,
777
- "single_word": false,
778
- "special": true
779
- },
780
- "32094": {
781
- "content": "<extra_id_5>",
782
- "lstrip": false,
783
- "normalized": false,
784
- "rstrip": false,
785
- "single_word": false,
786
- "special": true
787
- },
788
- "32095": {
789
- "content": "<extra_id_4>",
790
- "lstrip": false,
791
- "normalized": false,
792
- "rstrip": false,
793
- "single_word": false,
794
- "special": true
795
- },
796
- "32096": {
797
- "content": "<extra_id_3>",
798
- "lstrip": false,
799
- "normalized": false,
800
- "rstrip": false,
801
- "single_word": false,
802
- "special": true
803
- },
804
- "32097": {
805
- "content": "<extra_id_2>",
806
- "lstrip": false,
807
- "normalized": false,
808
- "rstrip": false,
809
- "single_word": false,
810
- "special": true
811
- },
812
- "32098": {
813
- "content": "<extra_id_1>",
814
- "lstrip": false,
815
- "normalized": false,
816
- "rstrip": false,
817
- "single_word": false,
818
- "special": true
819
- },
820
- "32099": {
821
- "content": "<extra_id_0>",
822
- "lstrip": false,
823
- "normalized": false,
824
- "rstrip": false,
825
- "single_word": false,
826
- "special": true
827
- }
828
- },
829
- "additional_special_tokens": [
830
- "<extra_id_0>",
831
- "<extra_id_1>",
832
- "<extra_id_2>",
833
- "<extra_id_3>",
834
- "<extra_id_4>",
835
- "<extra_id_5>",
836
- "<extra_id_6>",
837
- "<extra_id_7>",
838
- "<extra_id_8>",
839
- "<extra_id_9>",
840
- "<extra_id_10>",
841
- "<extra_id_11>",
842
- "<extra_id_12>",
843
- "<extra_id_13>",
844
- "<extra_id_14>",
845
- "<extra_id_15>",
846
- "<extra_id_16>",
847
- "<extra_id_17>",
848
- "<extra_id_18>",
849
- "<extra_id_19>",
850
- "<extra_id_20>",
851
- "<extra_id_21>",
852
- "<extra_id_22>",
853
- "<extra_id_23>",
854
- "<extra_id_24>",
855
- "<extra_id_25>",
856
- "<extra_id_26>",
857
- "<extra_id_27>",
858
- "<extra_id_28>",
859
- "<extra_id_29>",
860
- "<extra_id_30>",
861
- "<extra_id_31>",
862
- "<extra_id_32>",
863
- "<extra_id_33>",
864
- "<extra_id_34>",
865
- "<extra_id_35>",
866
- "<extra_id_36>",
867
- "<extra_id_37>",
868
- "<extra_id_38>",
869
- "<extra_id_39>",
870
- "<extra_id_40>",
871
- "<extra_id_41>",
872
- "<extra_id_42>",
873
- "<extra_id_43>",
874
- "<extra_id_44>",
875
- "<extra_id_45>",
876
- "<extra_id_46>",
877
- "<extra_id_47>",
878
- "<extra_id_48>",
879
- "<extra_id_49>",
880
- "<extra_id_50>",
881
- "<extra_id_51>",
882
- "<extra_id_52>",
883
- "<extra_id_53>",
884
- "<extra_id_54>",
885
- "<extra_id_55>",
886
- "<extra_id_56>",
887
- "<extra_id_57>",
888
- "<extra_id_58>",
889
- "<extra_id_59>",
890
- "<extra_id_60>",
891
- "<extra_id_61>",
892
- "<extra_id_62>",
893
- "<extra_id_63>",
894
- "<extra_id_64>",
895
- "<extra_id_65>",
896
- "<extra_id_66>",
897
- "<extra_id_67>",
898
- "<extra_id_68>",
899
- "<extra_id_69>",
900
- "<extra_id_70>",
901
- "<extra_id_71>",
902
- "<extra_id_72>",
903
- "<extra_id_73>",
904
- "<extra_id_74>",
905
- "<extra_id_75>",
906
- "<extra_id_76>",
907
- "<extra_id_77>",
908
- "<extra_id_78>",
909
- "<extra_id_79>",
910
- "<extra_id_80>",
911
- "<extra_id_81>",
912
- "<extra_id_82>",
913
- "<extra_id_83>",
914
- "<extra_id_84>",
915
- "<extra_id_85>",
916
- "<extra_id_86>",
917
- "<extra_id_87>",
918
- "<extra_id_88>",
919
- "<extra_id_89>",
920
- "<extra_id_90>",
921
- "<extra_id_91>",
922
- "<extra_id_92>",
923
- "<extra_id_93>",
924
- "<extra_id_94>",
925
- "<extra_id_95>",
926
- "<extra_id_96>",
927
- "<extra_id_97>",
928
- "<extra_id_98>",
929
- "<extra_id_99>"
930
- ],
931
- "clean_up_tokenization_spaces": true,
932
- "eos_token": "</s>",
933
- "extra_ids": 100,
934
- "legacy": true,
935
- "model_max_length": 512,
936
- "pad_token": "<pad>",
937
- "sp_model_kwargs": {},
938
- "tokenizer_class": "T5Tokenizer",
939
- "unk_token": "<unk>"
940
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/transformer/config.json DELETED
@@ -1,18 +0,0 @@
1
- {
2
- "_class_name": "FluxTransformer2DModel",
3
- "_diffusers_version": "0.30.1",
4
- "attention_head_dim": 128,
5
- "axes_dims_rope": [
6
- 16,
7
- 56,
8
- 56
9
- ],
10
- "guidance_embeds": true,
11
- "in_channels": 64,
12
- "joint_attention_dim": 4096,
13
- "num_attention_heads": 24,
14
- "num_layers": 19,
15
- "num_single_layers": 38,
16
- "patch_size": 1,
17
- "pooled_projection_dim": 768
18
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/transformer/diffusion_pytorch_model.safetensors.index.json DELETED
The diff for this file is too large to render. See raw diff
 
models/flux/vae/config.json DELETED
@@ -1,38 +0,0 @@
1
- {
2
- "_class_name": "AutoencoderKL",
3
- "_diffusers_version": "0.30.0.dev0",
4
- "_name_or_path": "../checkpoints/flux-dev",
5
- "act_fn": "silu",
6
- "block_out_channels": [
7
- 128,
8
- 256,
9
- 512,
10
- 512
11
- ],
12
- "down_block_types": [
13
- "DownEncoderBlock2D",
14
- "DownEncoderBlock2D",
15
- "DownEncoderBlock2D",
16
- "DownEncoderBlock2D"
17
- ],
18
- "force_upcast": true,
19
- "in_channels": 3,
20
- "latent_channels": 16,
21
- "latents_mean": null,
22
- "latents_std": null,
23
- "layers_per_block": 2,
24
- "mid_block_add_attention": true,
25
- "norm_num_groups": 32,
26
- "out_channels": 3,
27
- "sample_size": 1024,
28
- "scaling_factor": 0.3611,
29
- "shift_factor": 0.1159,
30
- "up_block_types": [
31
- "UpDecoderBlock2D",
32
- "UpDecoderBlock2D",
33
- "UpDecoderBlock2D",
34
- "UpDecoderBlock2D"
35
- ],
36
- "use_post_quant_conv": false,
37
- "use_quant_conv": false
38
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/flux/vae/diffusion_pytorch_model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5b59a26851551b67ae1fe58d32e76486e1e812def4696a4bea97f16604d40a3
3
- size 167666902
 
 
 
 
models/siglip2/README.md DELETED
@@ -1,102 +0,0 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - vision
5
- widget:
6
- - src: >-
7
- https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/bee.jpg
8
- candidate_labels: bee in the sky, bee on the flower
9
- example_title: Bee
10
- library_name: transformers
11
- pipeline_tag: zero-shot-image-classification
12
- ---
13
-
14
- # SigLIP 2 So400m
15
-
16
- [SigLIP 2](https://huggingface.co/papers/2502.14786) extends the pretraining objective of
17
- [SigLIP](https://huggingface.co/papers/2303.15343) with prior, independently developed techniques
18
- into a unified recipe, for improved semantic understanding, localization, and dense features.
19
-
20
- ## Intended uses
21
-
22
- You can use the raw model for tasks like zero-shot image classification and
23
- image-text retrieval, or as a vision encoder for VLMs (and other vision tasks).
24
-
25
- Here is how to use this model to perform zero-shot image classification:
26
-
27
- ```python
28
- from transformers import pipeline
29
-
30
- # load pipeline
31
- ckpt = "google/siglip2-so400m-patch16-512"
32
- image_classifier = pipeline(model=ckpt, task="zero-shot-image-classification")
33
-
34
- # load image and candidate labels
35
- url = "http://images.cocodataset.org/val2017/000000039769.jpg"
36
- candidate_labels = ["2 cats", "a plane", "a remote"]
37
-
38
- # run inference
39
- outputs = image_classifier(image, candidate_labels)
40
- print(outputs)
41
- ```
42
-
43
- You can encode an image using the Vision Tower like so:
44
-
45
- ```python
46
- import torch
47
- from transformers import AutoModel, AutoProcessor
48
- from transformers.image_utils import load_image
49
-
50
- # load the model and processor
51
- ckpt = "google/siglip2-so400m-patch16-512"
52
- model = AutoModel.from_pretrained(ckpt, device_map="auto").eval()
53
- processor = AutoProcessor.from_pretrained(ckpt)
54
-
55
- # load the image
56
- image = load_image("https://huggingface.co/datasets/merve/coco/resolve/main/val2017/000000000285.jpg")
57
- inputs = processor(images=[image], return_tensors="pt").to(model.device)
58
-
59
- # run infernece
60
- with torch.no_grad():
61
- image_embeddings = model.get_image_features(**inputs)
62
-
63
- print(image_embeddings.shape)
64
- ```
65
-
66
- For more code examples, we refer to the [siglip documentation](https://huggingface.co/transformers/main/model_doc/siglip.html#).
67
-
68
- ## Training procedure
69
-
70
- SigLIP 2 adds some clever training objectives on top of SigLIP:
71
-
72
- 1. Decoder loss
73
- 2. Global-local and masked prediction loss
74
- 3. Aspect ratio and resolution adaptibility
75
-
76
- ### Training data
77
-
78
- SigLIP 2 is pre-trained on the WebLI dataset [(Chen et al., 2023)](https://arxiv.org/abs/2209.06794).
79
-
80
- ### Compute
81
-
82
- The model was trained on up to 2048 TPU-v5e chips.
83
-
84
- ## Evaluation results
85
-
86
- Evaluation of SigLIP 2 is shown below (taken from the paper).
87
-
88
- ![Evaluation Table](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/sg2-blog/eval_table.png)
89
-
90
- ### BibTeX entry and citation info
91
-
92
- ```bibtex
93
- @misc{tschannen2025siglip2multilingualvisionlanguage,
94
- title={SigLIP 2: Multilingual Vision-Language Encoders with Improved Semantic Understanding, Localization, and Dense Features},
95
- author={Michael Tschannen and Alexey Gritsenko and Xiao Wang and Muhammad Ferjad Naeem and Ibrahim Alabdulmohsin and Nikhil Parthasarathy and Talfan Evans and Lucas Beyer and Ye Xia and Basil Mustafa and Olivier Hรฉnaff and Jeremiah Harmsen and Andreas Steiner and Xiaohua Zhai},
96
- year={2025},
97
- eprint={2502.14786},
98
- archivePrefix={arXiv},
99
- primaryClass={cs.CV},
100
- url={https://arxiv.org/abs/2502.14786},
101
- }
102
- ```
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/siglip2/config.json DELETED
@@ -1,22 +0,0 @@
1
- {
2
- "initializer_factor": 1.0,
3
- "model_type": "siglip",
4
- "text_config": {
5
- "hidden_size": 1152,
6
- "intermediate_size": 4304,
7
- "model_type": "siglip_text_model",
8
- "num_attention_heads": 16,
9
- "num_hidden_layers": 27,
10
- "projection_size": 1152,
11
- "vocab_size": 256000
12
- },
13
- "transformers_version": "4.49.0.dev0",
14
- "vision_config": {
15
- "hidden_size": 1152,
16
- "image_size": 512,
17
- "intermediate_size": 4304,
18
- "model_type": "siglip_vision_model",
19
- "num_attention_heads": 16,
20
- "num_hidden_layers": 27
21
- }
22
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/siglip2/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:a621bd212e1b3329b428595f9693217e19587afe826adf3e5c241a16392e8973
3
- size 4546331880
 
 
 
 
models/siglip2/preprocessor_config.json DELETED
@@ -1,24 +0,0 @@
1
- {
2
- "do_convert_rgb": null,
3
- "do_normalize": true,
4
- "do_rescale": true,
5
- "do_resize": true,
6
- "image_mean": [
7
- 0.5,
8
- 0.5,
9
- 0.5
10
- ],
11
- "image_processor_type": "SiglipImageProcessor",
12
- "image_std": [
13
- 0.5,
14
- 0.5,
15
- 0.5
16
- ],
17
- "processor_class": "SiglipProcessor",
18
- "resample": 2,
19
- "rescale_factor": 0.00392156862745098,
20
- "size": {
21
- "height": 512,
22
- "width": 512
23
- }
24
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/siglip2/put_them_to_siglip2 DELETED
@@ -1 +0,0 @@
1
- https://huggingface.co/google/siglip2-so400m-patch16-512
 
 
models/siglip2/special_tokens_map.json DELETED
@@ -1,34 +0,0 @@
1
- {
2
- "additional_special_tokens": [
3
- "<start_of_turn>",
4
- "<end_of_turn>"
5
- ],
6
- "bos_token": {
7
- "content": "<bos>",
8
- "lstrip": false,
9
- "normalized": false,
10
- "rstrip": false,
11
- "single_word": false
12
- },
13
- "eos_token": {
14
- "content": "<eos>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false
19
- },
20
- "pad_token": {
21
- "content": "<pad>",
22
- "lstrip": false,
23
- "normalized": false,
24
- "rstrip": false,
25
- "single_word": false
26
- },
27
- "unk_token": {
28
- "content": "<unk>",
29
- "lstrip": false,
30
- "normalized": false,
31
- "rstrip": false,
32
- "single_word": false
33
- }
34
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/siglip2/tokenizer.json DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb9140fae3ac5122c972d37adf83e1248471a38147ad76f8215c8872c6fd8322
3
- size 34363039
 
 
 
 
models/siglip2/tokenizer.model DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:61a7b147390c64585d6c3543dd6fc636906c9af3865a5548f27f31aee1d4c8e2
3
- size 4241003
 
 
 
 
models/siglip2/tokenizer_config.json DELETED
@@ -1,2020 +0,0 @@
1
- {
2
- "add_bos_token": false,
3
- "add_eos_token": true,
4
- "added_tokens_decoder": {
5
- "0": {
6
- "content": "<pad>",
7
- "lstrip": false,
8
- "normalized": false,
9
- "rstrip": false,
10
- "single_word": false,
11
- "special": true
12
- },
13
- "1": {
14
- "content": "<eos>",
15
- "lstrip": false,
16
- "normalized": false,
17
- "rstrip": false,
18
- "single_word": false,
19
- "special": true
20
- },
21
- "2": {
22
- "content": "<bos>",
23
- "lstrip": false,
24
- "normalized": false,
25
- "rstrip": false,
26
- "single_word": false,
27
- "special": true
28
- },
29
- "3": {
30
- "content": "<unk>",
31
- "lstrip": false,
32
- "normalized": false,
33
- "rstrip": false,
34
- "single_word": false,
35
- "special": true
36
- },
37
- "4": {
38
- "content": "<mask>",
39
- "lstrip": false,
40
- "normalized": false,
41
- "rstrip": false,
42
- "single_word": false,
43
- "special": false
44
- },
45
- "5": {
46
- "content": "<2mass>",
47
- "lstrip": false,
48
- "normalized": false,
49
- "rstrip": false,
50
- "single_word": false,
51
- "special": false
52
- },
53
- "6": {
54
- "content": "[@BOS@]",
55
- "lstrip": false,
56
- "normalized": false,
57
- "rstrip": false,
58
- "single_word": false,
59
- "special": false
60
- },
61
- "7": {
62
- "content": "<unused0>",
63
- "lstrip": false,
64
- "normalized": false,
65
- "rstrip": false,
66
- "single_word": false,
67
- "special": false
68
- },
69
- "8": {
70
- "content": "<unused1>",
71
- "lstrip": false,
72
- "normalized": false,
73
- "rstrip": false,
74
- "single_word": false,
75
- "special": false
76
- },
77
- "9": {
78
- "content": "<unused2>",
79
- "lstrip": false,
80
- "normalized": false,
81
- "rstrip": false,
82
- "single_word": false,
83
- "special": false
84
- },
85
- "10": {
86
- "content": "<unused3>",
87
- "lstrip": false,
88
- "normalized": false,
89
- "rstrip": false,
90
- "single_word": false,
91
- "special": false
92
- },
93
- "11": {
94
- "content": "<unused4>",
95
- "lstrip": false,
96
- "normalized": false,
97
- "rstrip": false,
98
- "single_word": false,
99
- "special": false
100
- },
101
- "12": {
102
- "content": "<unused5>",
103
- "lstrip": false,
104
- "normalized": false,
105
- "rstrip": false,
106
- "single_word": false,
107
- "special": false
108
- },
109
- "13": {
110
- "content": "<unused6>",
111
- "lstrip": false,
112
- "normalized": false,
113
- "rstrip": false,
114
- "single_word": false,
115
- "special": false
116
- },
117
- "14": {
118
- "content": "<unused7>",
119
- "lstrip": false,
120
- "normalized": false,
121
- "rstrip": false,
122
- "single_word": false,
123
- "special": false
124
- },
125
- "15": {
126
- "content": "<unused8>",
127
- "lstrip": false,
128
- "normalized": false,
129
- "rstrip": false,
130
- "single_word": false,
131
- "special": false
132
- },
133
- "16": {
134
- "content": "<unused9>",
135
- "lstrip": false,
136
- "normalized": false,
137
- "rstrip": false,
138
- "single_word": false,
139
- "special": false
140
- },
141
- "17": {
142
- "content": "<unused10>",
143
- "lstrip": false,
144
- "normalized": false,
145
- "rstrip": false,
146
- "single_word": false,
147
- "special": false
148
- },
149
- "18": {
150
- "content": "<unused11>",
151
- "lstrip": false,
152
- "normalized": false,
153
- "rstrip": false,
154
- "single_word": false,
155
- "special": false
156
- },
157
- "19": {
158
- "content": "<unused12>",
159
- "lstrip": false,
160
- "normalized": false,
161
- "rstrip": false,
162
- "single_word": false,
163
- "special": false
164
- },
165
- "20": {
166
- "content": "<unused13>",
167
- "lstrip": false,
168
- "normalized": false,
169
- "rstrip": false,
170
- "single_word": false,
171
- "special": false
172
- },
173
- "21": {
174
- "content": "<unused14>",
175
- "lstrip": false,
176
- "normalized": false,
177
- "rstrip": false,
178
- "single_word": false,
179
- "special": false
180
- },
181
- "22": {
182
- "content": "<unused15>",
183
- "lstrip": false,
184
- "normalized": false,
185
- "rstrip": false,
186
- "single_word": false,
187
- "special": false
188
- },
189
- "23": {
190
- "content": "<unused16>",
191
- "lstrip": false,
192
- "normalized": false,
193
- "rstrip": false,
194
- "single_word": false,
195
- "special": false
196
- },
197
- "24": {
198
- "content": "<unused17>",
199
- "lstrip": false,
200
- "normalized": false,
201
- "rstrip": false,
202
- "single_word": false,
203
- "special": false
204
- },
205
- "25": {
206
- "content": "<unused18>",
207
- "lstrip": false,
208
- "normalized": false,
209
- "rstrip": false,
210
- "single_word": false,
211
- "special": false
212
- },
213
- "26": {
214
- "content": "<unused19>",
215
- "lstrip": false,
216
- "normalized": false,
217
- "rstrip": false,
218
- "single_word": false,
219
- "special": false
220
- },
221
- "27": {
222
- "content": "<unused20>",
223
- "lstrip": false,
224
- "normalized": false,
225
- "rstrip": false,
226
- "single_word": false,
227
- "special": false
228
- },
229
- "28": {
230
- "content": "<unused21>",
231
- "lstrip": false,
232
- "normalized": false,
233
- "rstrip": false,
234
- "single_word": false,
235
- "special": false
236
- },
237
- "29": {
238
- "content": "<unused22>",
239
- "lstrip": false,
240
- "normalized": false,
241
- "rstrip": false,
242
- "single_word": false,
243
- "special": false
244
- },
245
- "30": {
246
- "content": "<unused23>",
247
- "lstrip": false,
248
- "normalized": false,
249
- "rstrip": false,
250
- "single_word": false,
251
- "special": false
252
- },
253
- "31": {
254
- "content": "<unused24>",
255
- "lstrip": false,
256
- "normalized": false,
257
- "rstrip": false,
258
- "single_word": false,
259
- "special": false
260
- },
261
- "32": {
262
- "content": "<unused25>",
263
- "lstrip": false,
264
- "normalized": false,
265
- "rstrip": false,
266
- "single_word": false,
267
- "special": false
268
- },
269
- "33": {
270
- "content": "<unused26>",
271
- "lstrip": false,
272
- "normalized": false,
273
- "rstrip": false,
274
- "single_word": false,
275
- "special": false
276
- },
277
- "34": {
278
- "content": "<unused27>",
279
- "lstrip": false,
280
- "normalized": false,
281
- "rstrip": false,
282
- "single_word": false,
283
- "special": false
284
- },
285
- "35": {
286
- "content": "<unused28>",
287
- "lstrip": false,
288
- "normalized": false,
289
- "rstrip": false,
290
- "single_word": false,
291
- "special": false
292
- },
293
- "36": {
294
- "content": "<unused29>",
295
- "lstrip": false,
296
- "normalized": false,
297
- "rstrip": false,
298
- "single_word": false,
299
- "special": false
300
- },
301
- "37": {
302
- "content": "<unused30>",
303
- "lstrip": false,
304
- "normalized": false,
305
- "rstrip": false,
306
- "single_word": false,
307
- "special": false
308
- },
309
- "38": {
310
- "content": "<unused31>",
311
- "lstrip": false,
312
- "normalized": false,
313
- "rstrip": false,
314
- "single_word": false,
315
- "special": false
316
- },
317
- "39": {
318
- "content": "<unused32>",
319
- "lstrip": false,
320
- "normalized": false,
321
- "rstrip": false,
322
- "single_word": false,
323
- "special": false
324
- },
325
- "40": {
326
- "content": "<unused33>",
327
- "lstrip": false,
328
- "normalized": false,
329
- "rstrip": false,
330
- "single_word": false,
331
- "special": false
332
- },
333
- "41": {
334
- "content": "<unused34>",
335
- "lstrip": false,
336
- "normalized": false,
337
- "rstrip": false,
338
- "single_word": false,
339
- "special": false
340
- },
341
- "42": {
342
- "content": "<unused35>",
343
- "lstrip": false,
344
- "normalized": false,
345
- "rstrip": false,
346
- "single_word": false,
347
- "special": false
348
- },
349
- "43": {
350
- "content": "<unused36>",
351
- "lstrip": false,
352
- "normalized": false,
353
- "rstrip": false,
354
- "single_word": false,
355
- "special": false
356
- },
357
- "44": {
358
- "content": "<unused37>",
359
- "lstrip": false,
360
- "normalized": false,
361
- "rstrip": false,
362
- "single_word": false,
363
- "special": false
364
- },
365
- "45": {
366
- "content": "<unused38>",
367
- "lstrip": false,
368
- "normalized": false,
369
- "rstrip": false,
370
- "single_word": false,
371
- "special": false
372
- },
373
- "46": {
374
- "content": "<unused39>",
375
- "lstrip": false,
376
- "normalized": false,
377
- "rstrip": false,
378
- "single_word": false,
379
- "special": false
380
- },
381
- "47": {
382
- "content": "<unused40>",
383
- "lstrip": false,
384
- "normalized": false,
385
- "rstrip": false,
386
- "single_word": false,
387
- "special": false
388
- },
389
- "48": {
390
- "content": "<unused41>",
391
- "lstrip": false,
392
- "normalized": false,
393
- "rstrip": false,
394
- "single_word": false,
395
- "special": false
396
- },
397
- "49": {
398
- "content": "<unused42>",
399
- "lstrip": false,
400
- "normalized": false,
401
- "rstrip": false,
402
- "single_word": false,
403
- "special": false
404
- },
405
- "50": {
406
- "content": "<unused43>",
407
- "lstrip": false,
408
- "normalized": false,
409
- "rstrip": false,
410
- "single_word": false,
411
- "special": false
412
- },
413
- "51": {
414
- "content": "<unused44>",
415
- "lstrip": false,
416
- "normalized": false,
417
- "rstrip": false,
418
- "single_word": false,
419
- "special": false
420
- },
421
- "52": {
422
- "content": "<unused45>",
423
- "lstrip": false,
424
- "normalized": false,
425
- "rstrip": false,
426
- "single_word": false,
427
- "special": false
428
- },
429
- "53": {
430
- "content": "<unused46>",
431
- "lstrip": false,
432
- "normalized": false,
433
- "rstrip": false,
434
- "single_word": false,
435
- "special": false
436
- },
437
- "54": {
438
- "content": "<unused47>",
439
- "lstrip": false,
440
- "normalized": false,
441
- "rstrip": false,
442
- "single_word": false,
443
- "special": false
444
- },
445
- "55": {
446
- "content": "<unused48>",
447
- "lstrip": false,
448
- "normalized": false,
449
- "rstrip": false,
450
- "single_word": false,
451
- "special": false
452
- },
453
- "56": {
454
- "content": "<unused49>",
455
- "lstrip": false,
456
- "normalized": false,
457
- "rstrip": false,
458
- "single_word": false,
459
- "special": false
460
- },
461
- "57": {
462
- "content": "<unused50>",
463
- "lstrip": false,
464
- "normalized": false,
465
- "rstrip": false,
466
- "single_word": false,
467
- "special": false
468
- },
469
- "58": {
470
- "content": "<unused51>",
471
- "lstrip": false,
472
- "normalized": false,
473
- "rstrip": false,
474
- "single_word": false,
475
- "special": false
476
- },
477
- "59": {
478
- "content": "<unused52>",
479
- "lstrip": false,
480
- "normalized": false,
481
- "rstrip": false,
482
- "single_word": false,
483
- "special": false
484
- },
485
- "60": {
486
- "content": "<unused53>",
487
- "lstrip": false,
488
- "normalized": false,
489
- "rstrip": false,
490
- "single_word": false,
491
- "special": false
492
- },
493
- "61": {
494
- "content": "<unused54>",
495
- "lstrip": false,
496
- "normalized": false,
497
- "rstrip": false,
498
- "single_word": false,
499
- "special": false
500
- },
501
- "62": {
502
- "content": "<unused55>",
503
- "lstrip": false,
504
- "normalized": false,
505
- "rstrip": false,
506
- "single_word": false,
507
- "special": false
508
- },
509
- "63": {
510
- "content": "<unused56>",
511
- "lstrip": false,
512
- "normalized": false,
513
- "rstrip": false,
514
- "single_word": false,
515
- "special": false
516
- },
517
- "64": {
518
- "content": "<unused57>",
519
- "lstrip": false,
520
- "normalized": false,
521
- "rstrip": false,
522
- "single_word": false,
523
- "special": false
524
- },
525
- "65": {
526
- "content": "<unused58>",
527
- "lstrip": false,
528
- "normalized": false,
529
- "rstrip": false,
530
- "single_word": false,
531
- "special": false
532
- },
533
- "66": {
534
- "content": "<unused59>",
535
- "lstrip": false,
536
- "normalized": false,
537
- "rstrip": false,
538
- "single_word": false,
539
- "special": false
540
- },
541
- "67": {
542
- "content": "<unused60>",
543
- "lstrip": false,
544
- "normalized": false,
545
- "rstrip": false,
546
- "single_word": false,
547
- "special": false
548
- },
549
- "68": {
550
- "content": "<unused61>",
551
- "lstrip": false,
552
- "normalized": false,
553
- "rstrip": false,
554
- "single_word": false,
555
- "special": false
556
- },
557
- "69": {
558
- "content": "<unused62>",
559
- "lstrip": false,
560
- "normalized": false,
561
- "rstrip": false,
562
- "single_word": false,
563
- "special": false
564
- },
565
- "70": {
566
- "content": "<unused63>",
567
- "lstrip": false,
568
- "normalized": false,
569
- "rstrip": false,
570
- "single_word": false,
571
- "special": false
572
- },
573
- "71": {
574
- "content": "<unused64>",
575
- "lstrip": false,
576
- "normalized": false,
577
- "rstrip": false,
578
- "single_word": false,
579
- "special": false
580
- },
581
- "72": {
582
- "content": "<unused65>",
583
- "lstrip": false,
584
- "normalized": false,
585
- "rstrip": false,
586
- "single_word": false,
587
- "special": false
588
- },
589
- "73": {
590
- "content": "<unused66>",
591
- "lstrip": false,
592
- "normalized": false,
593
- "rstrip": false,
594
- "single_word": false,
595
- "special": false
596
- },
597
- "74": {
598
- "content": "<unused67>",
599
- "lstrip": false,
600
- "normalized": false,
601
- "rstrip": false,
602
- "single_word": false,
603
- "special": false
604
- },
605
- "75": {
606
- "content": "<unused68>",
607
- "lstrip": false,
608
- "normalized": false,
609
- "rstrip": false,
610
- "single_word": false,
611
- "special": false
612
- },
613
- "76": {
614
- "content": "<unused69>",
615
- "lstrip": false,
616
- "normalized": false,
617
- "rstrip": false,
618
- "single_word": false,
619
- "special": false
620
- },
621
- "77": {
622
- "content": "<unused70>",
623
- "lstrip": false,
624
- "normalized": false,
625
- "rstrip": false,
626
- "single_word": false,
627
- "special": false
628
- },
629
- "78": {
630
- "content": "<unused71>",
631
- "lstrip": false,
632
- "normalized": false,
633
- "rstrip": false,
634
- "single_word": false,
635
- "special": false
636
- },
637
- "79": {
638
- "content": "<unused72>",
639
- "lstrip": false,
640
- "normalized": false,
641
- "rstrip": false,
642
- "single_word": false,
643
- "special": false
644
- },
645
- "80": {
646
- "content": "<unused73>",
647
- "lstrip": false,
648
- "normalized": false,
649
- "rstrip": false,
650
- "single_word": false,
651
- "special": false
652
- },
653
- "81": {
654
- "content": "<unused74>",
655
- "lstrip": false,
656
- "normalized": false,
657
- "rstrip": false,
658
- "single_word": false,
659
- "special": false
660
- },
661
- "82": {
662
- "content": "<unused75>",
663
- "lstrip": false,
664
- "normalized": false,
665
- "rstrip": false,
666
- "single_word": false,
667
- "special": false
668
- },
669
- "83": {
670
- "content": "<unused76>",
671
- "lstrip": false,
672
- "normalized": false,
673
- "rstrip": false,
674
- "single_word": false,
675
- "special": false
676
- },
677
- "84": {
678
- "content": "<unused77>",
679
- "lstrip": false,
680
- "normalized": false,
681
- "rstrip": false,
682
- "single_word": false,
683
- "special": false
684
- },
685
- "85": {
686
- "content": "<unused78>",
687
- "lstrip": false,
688
- "normalized": false,
689
- "rstrip": false,
690
- "single_word": false,
691
- "special": false
692
- },
693
- "86": {
694
- "content": "<unused79>",
695
- "lstrip": false,
696
- "normalized": false,
697
- "rstrip": false,
698
- "single_word": false,
699
- "special": false
700
- },
701
- "87": {
702
- "content": "<unused80>",
703
- "lstrip": false,
704
- "normalized": false,
705
- "rstrip": false,
706
- "single_word": false,
707
- "special": false
708
- },
709
- "88": {
710
- "content": "<unused81>",
711
- "lstrip": false,
712
- "normalized": false,
713
- "rstrip": false,
714
- "single_word": false,
715
- "special": false
716
- },
717
- "89": {
718
- "content": "<unused82>",
719
- "lstrip": false,
720
- "normalized": false,
721
- "rstrip": false,
722
- "single_word": false,
723
- "special": false
724
- },
725
- "90": {
726
- "content": "<unused83>",
727
- "lstrip": false,
728
- "normalized": false,
729
- "rstrip": false,
730
- "single_word": false,
731
- "special": false
732
- },
733
- "91": {
734
- "content": "<unused84>",
735
- "lstrip": false,
736
- "normalized": false,
737
- "rstrip": false,
738
- "single_word": false,
739
- "special": false
740
- },
741
- "92": {
742
- "content": "<unused85>",
743
- "lstrip": false,
744
- "normalized": false,
745
- "rstrip": false,
746
- "single_word": false,
747
- "special": false
748
- },
749
- "93": {
750
- "content": "<unused86>",
751
- "lstrip": false,
752
- "normalized": false,
753
- "rstrip": false,
754
- "single_word": false,
755
- "special": false
756
- },
757
- "94": {
758
- "content": "<unused87>",
759
- "lstrip": false,
760
- "normalized": false,
761
- "rstrip": false,
762
- "single_word": false,
763
- "special": false
764
- },
765
- "95": {
766
- "content": "<unused88>",
767
- "lstrip": false,
768
- "normalized": false,
769
- "rstrip": false,
770
- "single_word": false,
771
- "special": false
772
- },
773
- "96": {
774
- "content": "<unused89>",
775
- "lstrip": false,
776
- "normalized": false,
777
- "rstrip": false,
778
- "single_word": false,
779
- "special": false
780
- },
781
- "97": {
782
- "content": "<unused90>",
783
- "lstrip": false,
784
- "normalized": false,
785
- "rstrip": false,
786
- "single_word": false,
787
- "special": false
788
- },
789
- "98": {
790
- "content": "<unused91>",
791
- "lstrip": false,
792
- "normalized": false,
793
- "rstrip": false,
794
- "single_word": false,
795
- "special": false
796
- },
797
- "99": {
798
- "content": "<unused92>",
799
- "lstrip": false,
800
- "normalized": false,
801
- "rstrip": false,
802
- "single_word": false,
803
- "special": false
804
- },
805
- "100": {
806
- "content": "<unused93>",
807
- "lstrip": false,
808
- "normalized": false,
809
- "rstrip": false,
810
- "single_word": false,
811
- "special": false
812
- },
813
- "101": {
814
- "content": "<unused94>",
815
- "lstrip": false,
816
- "normalized": false,
817
- "rstrip": false,
818
- "single_word": false,
819
- "special": false
820
- },
821
- "102": {
822
- "content": "<unused95>",
823
- "lstrip": false,
824
- "normalized": false,
825
- "rstrip": false,
826
- "single_word": false,
827
- "special": false
828
- },
829
- "103": {
830
- "content": "<unused96>",
831
- "lstrip": false,
832
- "normalized": false,
833
- "rstrip": false,
834
- "single_word": false,
835
- "special": false
836
- },
837
- "104": {
838
- "content": "<unused97>",
839
- "lstrip": false,
840
- "normalized": false,
841
- "rstrip": false,
842
- "single_word": false,
843
- "special": false
844
- },
845
- "105": {
846
- "content": "<unused98>",
847
- "lstrip": false,
848
- "normalized": false,
849
- "rstrip": false,
850
- "single_word": false,
851
- "special": false
852
- },
853
- "106": {
854
- "content": "<start_of_turn>",
855
- "lstrip": false,
856
- "normalized": false,
857
- "rstrip": false,
858
- "single_word": false,
859
- "special": true
860
- },
861
- "107": {
862
- "content": "<end_of_turn>",
863
- "lstrip": false,
864
- "normalized": false,
865
- "rstrip": false,
866
- "single_word": false,
867
- "special": true
868
- },
869
- "108": {
870
- "content": "\n",
871
- "lstrip": false,
872
- "normalized": false,
873
- "rstrip": false,
874
- "single_word": false,
875
- "special": false
876
- },
877
- "109": {
878
- "content": "\n\n",
879
- "lstrip": false,
880
- "normalized": false,
881
- "rstrip": false,
882
- "single_word": false,
883
- "special": false
884
- },
885
- "110": {
886
- "content": "\n\n\n",
887
- "lstrip": false,
888
- "normalized": false,
889
- "rstrip": false,
890
- "single_word": false,
891
- "special": false
892
- },
893
- "111": {
894
- "content": "\n\n\n\n",
895
- "lstrip": false,
896
- "normalized": false,
897
- "rstrip": false,
898
- "single_word": false,
899
- "special": false
900
- },
901
- "112": {
902
- "content": "\n\n\n\n\n",
903
- "lstrip": false,
904
- "normalized": false,
905
- "rstrip": false,
906
- "single_word": false,
907
- "special": false
908
- },
909
- "113": {
910
- "content": "\n\n\n\n\n\n",
911
- "lstrip": false,
912
- "normalized": false,
913
- "rstrip": false,
914
- "single_word": false,
915
- "special": false
916
- },
917
- "114": {
918
- "content": "\n\n\n\n\n\n\n",
919
- "lstrip": false,
920
- "normalized": false,
921
- "rstrip": false,
922
- "single_word": false,
923
- "special": false
924
- },
925
- "115": {
926
- "content": "\n\n\n\n\n\n\n\n",
927
- "lstrip": false,
928
- "normalized": false,
929
- "rstrip": false,
930
- "single_word": false,
931
- "special": false
932
- },
933
- "116": {
934
- "content": "\n\n\n\n\n\n\n\n\n",
935
- "lstrip": false,
936
- "normalized": false,
937
- "rstrip": false,
938
- "single_word": false,
939
- "special": false
940
- },
941
- "117": {
942
- "content": "\n\n\n\n\n\n\n\n\n\n",
943
- "lstrip": false,
944
- "normalized": false,
945
- "rstrip": false,
946
- "single_word": false,
947
- "special": false
948
- },
949
- "118": {
950
- "content": "\n\n\n\n\n\n\n\n\n\n\n",
951
- "lstrip": false,
952
- "normalized": false,
953
- "rstrip": false,
954
- "single_word": false,
955
- "special": false
956
- },
957
- "119": {
958
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n",
959
- "lstrip": false,
960
- "normalized": false,
961
- "rstrip": false,
962
- "single_word": false,
963
- "special": false
964
- },
965
- "120": {
966
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n",
967
- "lstrip": false,
968
- "normalized": false,
969
- "rstrip": false,
970
- "single_word": false,
971
- "special": false
972
- },
973
- "121": {
974
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
975
- "lstrip": false,
976
- "normalized": false,
977
- "rstrip": false,
978
- "single_word": false,
979
- "special": false
980
- },
981
- "122": {
982
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
983
- "lstrip": false,
984
- "normalized": false,
985
- "rstrip": false,
986
- "single_word": false,
987
- "special": false
988
- },
989
- "123": {
990
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
991
- "lstrip": false,
992
- "normalized": false,
993
- "rstrip": false,
994
- "single_word": false,
995
- "special": false
996
- },
997
- "124": {
998
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
999
- "lstrip": false,
1000
- "normalized": false,
1001
- "rstrip": false,
1002
- "single_word": false,
1003
- "special": false
1004
- },
1005
- "125": {
1006
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1007
- "lstrip": false,
1008
- "normalized": false,
1009
- "rstrip": false,
1010
- "single_word": false,
1011
- "special": false
1012
- },
1013
- "126": {
1014
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1015
- "lstrip": false,
1016
- "normalized": false,
1017
- "rstrip": false,
1018
- "single_word": false,
1019
- "special": false
1020
- },
1021
- "127": {
1022
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1023
- "lstrip": false,
1024
- "normalized": false,
1025
- "rstrip": false,
1026
- "single_word": false,
1027
- "special": false
1028
- },
1029
- "128": {
1030
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1031
- "lstrip": false,
1032
- "normalized": false,
1033
- "rstrip": false,
1034
- "single_word": false,
1035
- "special": false
1036
- },
1037
- "129": {
1038
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1039
- "lstrip": false,
1040
- "normalized": false,
1041
- "rstrip": false,
1042
- "single_word": false,
1043
- "special": false
1044
- },
1045
- "130": {
1046
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1047
- "lstrip": false,
1048
- "normalized": false,
1049
- "rstrip": false,
1050
- "single_word": false,
1051
- "special": false
1052
- },
1053
- "131": {
1054
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1055
- "lstrip": false,
1056
- "normalized": false,
1057
- "rstrip": false,
1058
- "single_word": false,
1059
- "special": false
1060
- },
1061
- "132": {
1062
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1063
- "lstrip": false,
1064
- "normalized": false,
1065
- "rstrip": false,
1066
- "single_word": false,
1067
- "special": false
1068
- },
1069
- "133": {
1070
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1071
- "lstrip": false,
1072
- "normalized": false,
1073
- "rstrip": false,
1074
- "single_word": false,
1075
- "special": false
1076
- },
1077
- "134": {
1078
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1079
- "lstrip": false,
1080
- "normalized": false,
1081
- "rstrip": false,
1082
- "single_word": false,
1083
- "special": false
1084
- },
1085
- "135": {
1086
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1087
- "lstrip": false,
1088
- "normalized": false,
1089
- "rstrip": false,
1090
- "single_word": false,
1091
- "special": false
1092
- },
1093
- "136": {
1094
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1095
- "lstrip": false,
1096
- "normalized": false,
1097
- "rstrip": false,
1098
- "single_word": false,
1099
- "special": false
1100
- },
1101
- "137": {
1102
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1103
- "lstrip": false,
1104
- "normalized": false,
1105
- "rstrip": false,
1106
- "single_word": false,
1107
- "special": false
1108
- },
1109
- "138": {
1110
- "content": "\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n\n",
1111
- "lstrip": false,
1112
- "normalized": false,
1113
- "rstrip": false,
1114
- "single_word": false,
1115
- "special": false
1116
- },
1117
- "139": {
1118
- "content": "โ–โ–",
1119
- "lstrip": false,
1120
- "normalized": false,
1121
- "rstrip": false,
1122
- "single_word": false,
1123
- "special": false
1124
- },
1125
- "140": {
1126
- "content": "โ–โ–โ–",
1127
- "lstrip": false,
1128
- "normalized": false,
1129
- "rstrip": false,
1130
- "single_word": false,
1131
- "special": false
1132
- },
1133
- "141": {
1134
- "content": "โ–โ–โ–โ–",
1135
- "lstrip": false,
1136
- "normalized": false,
1137
- "rstrip": false,
1138
- "single_word": false,
1139
- "special": false
1140
- },
1141
- "142": {
1142
- "content": "โ–โ–โ–โ–โ–",
1143
- "lstrip": false,
1144
- "normalized": false,
1145
- "rstrip": false,
1146
- "single_word": false,
1147
- "special": false
1148
- },
1149
- "143": {
1150
- "content": "โ–โ–โ–โ–โ–โ–",
1151
- "lstrip": false,
1152
- "normalized": false,
1153
- "rstrip": false,
1154
- "single_word": false,
1155
- "special": false
1156
- },
1157
- "144": {
1158
- "content": "โ–โ–โ–โ–โ–โ–โ–",
1159
- "lstrip": false,
1160
- "normalized": false,
1161
- "rstrip": false,
1162
- "single_word": false,
1163
- "special": false
1164
- },
1165
- "145": {
1166
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–",
1167
- "lstrip": false,
1168
- "normalized": false,
1169
- "rstrip": false,
1170
- "single_word": false,
1171
- "special": false
1172
- },
1173
- "146": {
1174
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1175
- "lstrip": false,
1176
- "normalized": false,
1177
- "rstrip": false,
1178
- "single_word": false,
1179
- "special": false
1180
- },
1181
- "147": {
1182
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1183
- "lstrip": false,
1184
- "normalized": false,
1185
- "rstrip": false,
1186
- "single_word": false,
1187
- "special": false
1188
- },
1189
- "148": {
1190
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1191
- "lstrip": false,
1192
- "normalized": false,
1193
- "rstrip": false,
1194
- "single_word": false,
1195
- "special": false
1196
- },
1197
- "149": {
1198
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1199
- "lstrip": false,
1200
- "normalized": false,
1201
- "rstrip": false,
1202
- "single_word": false,
1203
- "special": false
1204
- },
1205
- "150": {
1206
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1207
- "lstrip": false,
1208
- "normalized": false,
1209
- "rstrip": false,
1210
- "single_word": false,
1211
- "special": false
1212
- },
1213
- "151": {
1214
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1215
- "lstrip": false,
1216
- "normalized": false,
1217
- "rstrip": false,
1218
- "single_word": false,
1219
- "special": false
1220
- },
1221
- "152": {
1222
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1223
- "lstrip": false,
1224
- "normalized": false,
1225
- "rstrip": false,
1226
- "single_word": false,
1227
- "special": false
1228
- },
1229
- "153": {
1230
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1231
- "lstrip": false,
1232
- "normalized": false,
1233
- "rstrip": false,
1234
- "single_word": false,
1235
- "special": false
1236
- },
1237
- "154": {
1238
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1239
- "lstrip": false,
1240
- "normalized": false,
1241
- "rstrip": false,
1242
- "single_word": false,
1243
- "special": false
1244
- },
1245
- "155": {
1246
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1247
- "lstrip": false,
1248
- "normalized": false,
1249
- "rstrip": false,
1250
- "single_word": false,
1251
- "special": false
1252
- },
1253
- "156": {
1254
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1255
- "lstrip": false,
1256
- "normalized": false,
1257
- "rstrip": false,
1258
- "single_word": false,
1259
- "special": false
1260
- },
1261
- "157": {
1262
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1263
- "lstrip": false,
1264
- "normalized": false,
1265
- "rstrip": false,
1266
- "single_word": false,
1267
- "special": false
1268
- },
1269
- "158": {
1270
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1271
- "lstrip": false,
1272
- "normalized": false,
1273
- "rstrip": false,
1274
- "single_word": false,
1275
- "special": false
1276
- },
1277
- "159": {
1278
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1279
- "lstrip": false,
1280
- "normalized": false,
1281
- "rstrip": false,
1282
- "single_word": false,
1283
- "special": false
1284
- },
1285
- "160": {
1286
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1287
- "lstrip": false,
1288
- "normalized": false,
1289
- "rstrip": false,
1290
- "single_word": false,
1291
- "special": false
1292
- },
1293
- "161": {
1294
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1295
- "lstrip": false,
1296
- "normalized": false,
1297
- "rstrip": false,
1298
- "single_word": false,
1299
- "special": false
1300
- },
1301
- "162": {
1302
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1303
- "lstrip": false,
1304
- "normalized": false,
1305
- "rstrip": false,
1306
- "single_word": false,
1307
- "special": false
1308
- },
1309
- "163": {
1310
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1311
- "lstrip": false,
1312
- "normalized": false,
1313
- "rstrip": false,
1314
- "single_word": false,
1315
- "special": false
1316
- },
1317
- "164": {
1318
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1319
- "lstrip": false,
1320
- "normalized": false,
1321
- "rstrip": false,
1322
- "single_word": false,
1323
- "special": false
1324
- },
1325
- "165": {
1326
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1327
- "lstrip": false,
1328
- "normalized": false,
1329
- "rstrip": false,
1330
- "single_word": false,
1331
- "special": false
1332
- },
1333
- "166": {
1334
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1335
- "lstrip": false,
1336
- "normalized": false,
1337
- "rstrip": false,
1338
- "single_word": false,
1339
- "special": false
1340
- },
1341
- "167": {
1342
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1343
- "lstrip": false,
1344
- "normalized": false,
1345
- "rstrip": false,
1346
- "single_word": false,
1347
- "special": false
1348
- },
1349
- "168": {
1350
- "content": "โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–โ–",
1351
- "lstrip": false,
1352
- "normalized": false,
1353
- "rstrip": false,
1354
- "single_word": false,
1355
- "special": false
1356
- },
1357
- "169": {
1358
- "content": "<table>",
1359
- "lstrip": false,
1360
- "normalized": false,
1361
- "rstrip": false,
1362
- "single_word": false,
1363
- "special": false
1364
- },
1365
- "170": {
1366
- "content": "<caption>",
1367
- "lstrip": false,
1368
- "normalized": false,
1369
- "rstrip": false,
1370
- "single_word": false,
1371
- "special": false
1372
- },
1373
- "171": {
1374
- "content": "<thead>",
1375
- "lstrip": false,
1376
- "normalized": false,
1377
- "rstrip": false,
1378
- "single_word": false,
1379
- "special": false
1380
- },
1381
- "172": {
1382
- "content": "<tbody>",
1383
- "lstrip": false,
1384
- "normalized": false,
1385
- "rstrip": false,
1386
- "single_word": false,
1387
- "special": false
1388
- },
1389
- "173": {
1390
- "content": "<tfoot>",
1391
- "lstrip": false,
1392
- "normalized": false,
1393
- "rstrip": false,
1394
- "single_word": false,
1395
- "special": false
1396
- },
1397
- "174": {
1398
- "content": "<tr>",
1399
- "lstrip": false,
1400
- "normalized": false,
1401
- "rstrip": false,
1402
- "single_word": false,
1403
- "special": false
1404
- },
1405
- "175": {
1406
- "content": "<th>",
1407
- "lstrip": false,
1408
- "normalized": false,
1409
- "rstrip": false,
1410
- "single_word": false,
1411
- "special": false
1412
- },
1413
- "176": {
1414
- "content": "<td>",
1415
- "lstrip": false,
1416
- "normalized": false,
1417
- "rstrip": false,
1418
- "single_word": false,
1419
- "special": false
1420
- },
1421
- "177": {
1422
- "content": "</table>",
1423
- "lstrip": false,
1424
- "normalized": false,
1425
- "rstrip": false,
1426
- "single_word": false,
1427
- "special": false
1428
- },
1429
- "178": {
1430
- "content": "</caption>",
1431
- "lstrip": false,
1432
- "normalized": false,
1433
- "rstrip": false,
1434
- "single_word": false,
1435
- "special": false
1436
- },
1437
- "179": {
1438
- "content": "</thead>",
1439
- "lstrip": false,
1440
- "normalized": false,
1441
- "rstrip": false,
1442
- "single_word": false,
1443
- "special": false
1444
- },
1445
- "180": {
1446
- "content": "</tbody>",
1447
- "lstrip": false,
1448
- "normalized": false,
1449
- "rstrip": false,
1450
- "single_word": false,
1451
- "special": false
1452
- },
1453
- "181": {
1454
- "content": "</tfoot>",
1455
- "lstrip": false,
1456
- "normalized": false,
1457
- "rstrip": false,
1458
- "single_word": false,
1459
- "special": false
1460
- },
1461
- "182": {
1462
- "content": "</tr>",
1463
- "lstrip": false,
1464
- "normalized": false,
1465
- "rstrip": false,
1466
- "single_word": false,
1467
- "special": false
1468
- },
1469
- "183": {
1470
- "content": "</th>",
1471
- "lstrip": false,
1472
- "normalized": false,
1473
- "rstrip": false,
1474
- "single_word": false,
1475
- "special": false
1476
- },
1477
- "184": {
1478
- "content": "</td>",
1479
- "lstrip": false,
1480
- "normalized": false,
1481
- "rstrip": false,
1482
- "single_word": false,
1483
- "special": false
1484
- },
1485
- "185": {
1486
- "content": "<h1>",
1487
- "lstrip": false,
1488
- "normalized": false,
1489
- "rstrip": false,
1490
- "single_word": false,
1491
- "special": false
1492
- },
1493
- "186": {
1494
- "content": "<h2>",
1495
- "lstrip": false,
1496
- "normalized": false,
1497
- "rstrip": false,
1498
- "single_word": false,
1499
- "special": false
1500
- },
1501
- "187": {
1502
- "content": "<h3>",
1503
- "lstrip": false,
1504
- "normalized": false,
1505
- "rstrip": false,
1506
- "single_word": false,
1507
- "special": false
1508
- },
1509
- "188": {
1510
- "content": "<h4>",
1511
- "lstrip": false,
1512
- "normalized": false,
1513
- "rstrip": false,
1514
- "single_word": false,
1515
- "special": false
1516
- },
1517
- "189": {
1518
- "content": "<h5>",
1519
- "lstrip": false,
1520
- "normalized": false,
1521
- "rstrip": false,
1522
- "single_word": false,
1523
- "special": false
1524
- },
1525
- "190": {
1526
- "content": "<h6>",
1527
- "lstrip": false,
1528
- "normalized": false,
1529
- "rstrip": false,
1530
- "single_word": false,
1531
- "special": false
1532
- },
1533
- "191": {
1534
- "content": "<blockquote>",
1535
- "lstrip": false,
1536
- "normalized": false,
1537
- "rstrip": false,
1538
- "single_word": false,
1539
- "special": false
1540
- },
1541
- "192": {
1542
- "content": "</h1>",
1543
- "lstrip": false,
1544
- "normalized": false,
1545
- "rstrip": false,
1546
- "single_word": false,
1547
- "special": false
1548
- },
1549
- "193": {
1550
- "content": "</h2>",
1551
- "lstrip": false,
1552
- "normalized": false,
1553
- "rstrip": false,
1554
- "single_word": false,
1555
- "special": false
1556
- },
1557
- "194": {
1558
- "content": "</h3>",
1559
- "lstrip": false,
1560
- "normalized": false,
1561
- "rstrip": false,
1562
- "single_word": false,
1563
- "special": false
1564
- },
1565
- "195": {
1566
- "content": "</h4>",
1567
- "lstrip": false,
1568
- "normalized": false,
1569
- "rstrip": false,
1570
- "single_word": false,
1571
- "special": false
1572
- },
1573
- "196": {
1574
- "content": "</h5>",
1575
- "lstrip": false,
1576
- "normalized": false,
1577
- "rstrip": false,
1578
- "single_word": false,
1579
- "special": false
1580
- },
1581
- "197": {
1582
- "content": "</h6>",
1583
- "lstrip": false,
1584
- "normalized": false,
1585
- "rstrip": false,
1586
- "single_word": false,
1587
- "special": false
1588
- },
1589
- "198": {
1590
- "content": "</blockquote>",
1591
- "lstrip": false,
1592
- "normalized": false,
1593
- "rstrip": false,
1594
- "single_word": false,
1595
- "special": false
1596
- },
1597
- "199": {
1598
- "content": "<strong>",
1599
- "lstrip": false,
1600
- "normalized": false,
1601
- "rstrip": false,
1602
- "single_word": false,
1603
- "special": false
1604
- },
1605
- "200": {
1606
- "content": "<em>",
1607
- "lstrip": false,
1608
- "normalized": false,
1609
- "rstrip": false,
1610
- "single_word": false,
1611
- "special": false
1612
- },
1613
- "201": {
1614
- "content": "<b>",
1615
- "lstrip": false,
1616
- "normalized": false,
1617
- "rstrip": false,
1618
- "single_word": false,
1619
- "special": false
1620
- },
1621
- "202": {
1622
- "content": "<i>",
1623
- "lstrip": false,
1624
- "normalized": false,
1625
- "rstrip": false,
1626
- "single_word": false,
1627
- "special": false
1628
- },
1629
- "203": {
1630
- "content": "<u>",
1631
- "lstrip": false,
1632
- "normalized": false,
1633
- "rstrip": false,
1634
- "single_word": false,
1635
- "special": false
1636
- },
1637
- "204": {
1638
- "content": "<s>",
1639
- "lstrip": false,
1640
- "normalized": false,
1641
- "rstrip": false,
1642
- "single_word": false,
1643
- "special": false
1644
- },
1645
- "205": {
1646
- "content": "<sub>",
1647
- "lstrip": false,
1648
- "normalized": false,
1649
- "rstrip": false,
1650
- "single_word": false,
1651
- "special": false
1652
- },
1653
- "206": {
1654
- "content": "<sup>",
1655
- "lstrip": false,
1656
- "normalized": false,
1657
- "rstrip": false,
1658
- "single_word": false,
1659
- "special": false
1660
- },
1661
- "207": {
1662
- "content": "<code>",
1663
- "lstrip": false,
1664
- "normalized": false,
1665
- "rstrip": false,
1666
- "single_word": false,
1667
- "special": false
1668
- },
1669
- "208": {
1670
- "content": "</strong>",
1671
- "lstrip": false,
1672
- "normalized": false,
1673
- "rstrip": false,
1674
- "single_word": false,
1675
- "special": false
1676
- },
1677
- "209": {
1678
- "content": "</em>",
1679
- "lstrip": false,
1680
- "normalized": false,
1681
- "rstrip": false,
1682
- "single_word": false,
1683
- "special": false
1684
- },
1685
- "210": {
1686
- "content": "</b>",
1687
- "lstrip": false,
1688
- "normalized": false,
1689
- "rstrip": false,
1690
- "single_word": false,
1691
- "special": false
1692
- },
1693
- "211": {
1694
- "content": "</i>",
1695
- "lstrip": false,
1696
- "normalized": false,
1697
- "rstrip": false,
1698
- "single_word": false,
1699
- "special": false
1700
- },
1701
- "212": {
1702
- "content": "</u>",
1703
- "lstrip": false,
1704
- "normalized": false,
1705
- "rstrip": false,
1706
- "single_word": false,
1707
- "special": false
1708
- },
1709
- "213": {
1710
- "content": "</s>",
1711
- "lstrip": false,
1712
- "normalized": false,
1713
- "rstrip": false,
1714
- "single_word": false,
1715
- "special": false
1716
- },
1717
- "214": {
1718
- "content": "</sub>",
1719
- "lstrip": false,
1720
- "normalized": false,
1721
- "rstrip": false,
1722
- "single_word": false,
1723
- "special": false
1724
- },
1725
- "215": {
1726
- "content": "</sup>",
1727
- "lstrip": false,
1728
- "normalized": false,
1729
- "rstrip": false,
1730
- "single_word": false,
1731
- "special": false
1732
- },
1733
- "216": {
1734
- "content": "</code>",
1735
- "lstrip": false,
1736
- "normalized": false,
1737
- "rstrip": false,
1738
- "single_word": false,
1739
- "special": false
1740
- },
1741
- "255968": {
1742
- "content": "[toxicity=0]",
1743
- "lstrip": false,
1744
- "normalized": false,
1745
- "rstrip": false,
1746
- "single_word": false,
1747
- "special": false
1748
- },
1749
- "255969": {
1750
- "content": "\t\t",
1751
- "lstrip": false,
1752
- "normalized": false,
1753
- "rstrip": false,
1754
- "single_word": false,
1755
- "special": false
1756
- },
1757
- "255970": {
1758
- "content": "\t\t\t",
1759
- "lstrip": false,
1760
- "normalized": false,
1761
- "rstrip": false,
1762
- "single_word": false,
1763
- "special": false
1764
- },
1765
- "255971": {
1766
- "content": "\t\t\t\t",
1767
- "lstrip": false,
1768
- "normalized": false,
1769
- "rstrip": false,
1770
- "single_word": false,
1771
- "special": false
1772
- },
1773
- "255972": {
1774
- "content": "\t\t\t\t\t",
1775
- "lstrip": false,
1776
- "normalized": false,
1777
- "rstrip": false,
1778
- "single_word": false,
1779
- "special": false
1780
- },
1781
- "255973": {
1782
- "content": "\t\t\t\t\t\t",
1783
- "lstrip": false,
1784
- "normalized": false,
1785
- "rstrip": false,
1786
- "single_word": false,
1787
- "special": false
1788
- },
1789
- "255974": {
1790
- "content": "\t\t\t\t\t\t\t",
1791
- "lstrip": false,
1792
- "normalized": false,
1793
- "rstrip": false,
1794
- "single_word": false,
1795
- "special": false
1796
- },
1797
- "255975": {
1798
- "content": "\t\t\t\t\t\t\t\t",
1799
- "lstrip": false,
1800
- "normalized": false,
1801
- "rstrip": false,
1802
- "single_word": false,
1803
- "special": false
1804
- },
1805
- "255976": {
1806
- "content": "\t\t\t\t\t\t\t\t\t",
1807
- "lstrip": false,
1808
- "normalized": false,
1809
- "rstrip": false,
1810
- "single_word": false,
1811
- "special": false
1812
- },
1813
- "255977": {
1814
- "content": "\t\t\t\t\t\t\t\t\t\t",
1815
- "lstrip": false,
1816
- "normalized": false,
1817
- "rstrip": false,
1818
- "single_word": false,
1819
- "special": false
1820
- },
1821
- "255978": {
1822
- "content": "\t\t\t\t\t\t\t\t\t\t\t",
1823
- "lstrip": false,
1824
- "normalized": false,
1825
- "rstrip": false,
1826
- "single_word": false,
1827
- "special": false
1828
- },
1829
- "255979": {
1830
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t",
1831
- "lstrip": false,
1832
- "normalized": false,
1833
- "rstrip": false,
1834
- "single_word": false,
1835
- "special": false
1836
- },
1837
- "255980": {
1838
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t",
1839
- "lstrip": false,
1840
- "normalized": false,
1841
- "rstrip": false,
1842
- "single_word": false,
1843
- "special": false
1844
- },
1845
- "255981": {
1846
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1847
- "lstrip": false,
1848
- "normalized": false,
1849
- "rstrip": false,
1850
- "single_word": false,
1851
- "special": false
1852
- },
1853
- "255982": {
1854
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1855
- "lstrip": false,
1856
- "normalized": false,
1857
- "rstrip": false,
1858
- "single_word": false,
1859
- "special": false
1860
- },
1861
- "255983": {
1862
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1863
- "lstrip": false,
1864
- "normalized": false,
1865
- "rstrip": false,
1866
- "single_word": false,
1867
- "special": false
1868
- },
1869
- "255984": {
1870
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1871
- "lstrip": false,
1872
- "normalized": false,
1873
- "rstrip": false,
1874
- "single_word": false,
1875
- "special": false
1876
- },
1877
- "255985": {
1878
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1879
- "lstrip": false,
1880
- "normalized": false,
1881
- "rstrip": false,
1882
- "single_word": false,
1883
- "special": false
1884
- },
1885
- "255986": {
1886
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1887
- "lstrip": false,
1888
- "normalized": false,
1889
- "rstrip": false,
1890
- "single_word": false,
1891
- "special": false
1892
- },
1893
- "255987": {
1894
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1895
- "lstrip": false,
1896
- "normalized": false,
1897
- "rstrip": false,
1898
- "single_word": false,
1899
- "special": false
1900
- },
1901
- "255988": {
1902
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1903
- "lstrip": false,
1904
- "normalized": false,
1905
- "rstrip": false,
1906
- "single_word": false,
1907
- "special": false
1908
- },
1909
- "255989": {
1910
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1911
- "lstrip": false,
1912
- "normalized": false,
1913
- "rstrip": false,
1914
- "single_word": false,
1915
- "special": false
1916
- },
1917
- "255990": {
1918
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1919
- "lstrip": false,
1920
- "normalized": false,
1921
- "rstrip": false,
1922
- "single_word": false,
1923
- "special": false
1924
- },
1925
- "255991": {
1926
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1927
- "lstrip": false,
1928
- "normalized": false,
1929
- "rstrip": false,
1930
- "single_word": false,
1931
- "special": false
1932
- },
1933
- "255992": {
1934
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1935
- "lstrip": false,
1936
- "normalized": false,
1937
- "rstrip": false,
1938
- "single_word": false,
1939
- "special": false
1940
- },
1941
- "255993": {
1942
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1943
- "lstrip": false,
1944
- "normalized": false,
1945
- "rstrip": false,
1946
- "single_word": false,
1947
- "special": false
1948
- },
1949
- "255994": {
1950
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1951
- "lstrip": false,
1952
- "normalized": false,
1953
- "rstrip": false,
1954
- "single_word": false,
1955
- "special": false
1956
- },
1957
- "255995": {
1958
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1959
- "lstrip": false,
1960
- "normalized": false,
1961
- "rstrip": false,
1962
- "single_word": false,
1963
- "special": false
1964
- },
1965
- "255996": {
1966
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1967
- "lstrip": false,
1968
- "normalized": false,
1969
- "rstrip": false,
1970
- "single_word": false,
1971
- "special": false
1972
- },
1973
- "255997": {
1974
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1975
- "lstrip": false,
1976
- "normalized": false,
1977
- "rstrip": false,
1978
- "single_word": false,
1979
- "special": false
1980
- },
1981
- "255998": {
1982
- "content": "\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t\t",
1983
- "lstrip": false,
1984
- "normalized": false,
1985
- "rstrip": false,
1986
- "single_word": false,
1987
- "special": false
1988
- },
1989
- "255999": {
1990
- "content": "<unused99>",
1991
- "lstrip": false,
1992
- "normalized": false,
1993
- "rstrip": false,
1994
- "single_word": false,
1995
- "special": false
1996
- }
1997
- },
1998
- "additional_special_tokens": [
1999
- "<start_of_turn>",
2000
- "<end_of_turn>"
2001
- ],
2002
- "bos_token": "<bos>",
2003
- "chat_template": "{{ bos_token }}{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}",
2004
- "clean_up_tokenization_spaces": false,
2005
- "do_lower_case": true,
2006
- "eos_token": "<eos>",
2007
- "extra_special_tokens": {},
2008
- "model_input_names": [
2009
- "input_ids"
2010
- ],
2011
- "model_max_length": 1000000000000000019884624838656,
2012
- "pad_token": "<pad>",
2013
- "padding_side": "right",
2014
- "processor_class": "SiglipProcessor",
2015
- "sp_model_kwargs": {},
2016
- "spaces_between_special_tokens": false,
2017
- "tokenizer_class": "GemmaTokenizer",
2018
- "unk_token": "<unk>",
2019
- "use_default_system_prompt": false
2020
- }