ArthurZ HF Staff commited on
Commit
497cd30
·
1 Parent(s): 41a5f5e

Update checkpoint for transformers>=4.29

Browse files

Following the merge of [a PR](https://github.com/huggingface/transformers/pull/24310) in `transformers` it appeared that this model was not properly converted. This PR will fix the inference and was tested using the following script:
```python
>>> from transformers import MarianModel, MarianMTModel
>>> tokenizer = AutoTokenizer.from_pretrained(Helsinki-NLP/opus-mt-tc-big-en-ar)
>>> inputs = tokenizer(Hey! Let's learn together, return_tensors=pt, padding=True)
>>> model = MarianMTModel.from_pretrained(Helsinki-NLP/opus-mt-tc-big-en-ar)
>>> print(tokenizer.batch_decode(model.generate(**inputs)))
['<pad> لنتعلّم معاً</s>']
```

config.json CHANGED
@@ -7,7 +7,7 @@
7
  "attention_dropout": 0.0,
8
  "bad_words_ids": [
9
  [
10
- 35697
11
  ]
12
  ],
13
  "bos_token_id": 0,
@@ -17,15 +17,15 @@
17
  "decoder_ffn_dim": 4096,
18
  "decoder_layerdrop": 0.0,
19
  "decoder_layers": 6,
20
- "decoder_start_token_id": 35697,
21
- "decoder_vocab_size": 35698,
22
  "dropout": 0.1,
23
  "encoder_attention_heads": 16,
24
  "encoder_ffn_dim": 4096,
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 6,
27
- "eos_token_id": 26162,
28
- "forced_eos_token_id": 26162,
29
  "init_std": 0.02,
30
  "is_encoder_decoder": true,
31
  "max_length": 512,
@@ -34,12 +34,12 @@
34
  "normalize_embedding": false,
35
  "num_beams": 4,
36
  "num_hidden_layers": 6,
37
- "pad_token_id": 35697,
38
  "scale_embedding": true,
39
  "share_encoder_decoder_embeddings": true,
40
  "static_position_embeddings": true,
41
  "torch_dtype": "float16",
42
  "transformers_version": "4.34.0.dev0",
43
  "use_cache": true,
44
- "vocab_size": 35698
45
  }
 
7
  "attention_dropout": 0.0,
8
  "bad_words_ids": [
9
  [
10
+ 61246
11
  ]
12
  ],
13
  "bos_token_id": 0,
 
17
  "decoder_ffn_dim": 4096,
18
  "decoder_layerdrop": 0.0,
19
  "decoder_layers": 6,
20
+ "decoder_start_token_id": 61246,
21
+ "decoder_vocab_size": 61247,
22
  "dropout": 0.1,
23
  "encoder_attention_heads": 16,
24
  "encoder_ffn_dim": 4096,
25
  "encoder_layerdrop": 0.0,
26
  "encoder_layers": 6,
27
+ "eos_token_id": 25897,
28
+ "forced_eos_token_id": 25897,
29
  "init_std": 0.02,
30
  "is_encoder_decoder": true,
31
  "max_length": 512,
 
34
  "normalize_embedding": false,
35
  "num_beams": 4,
36
  "num_hidden_layers": 6,
37
+ "pad_token_id": 61246,
38
  "scale_embedding": true,
39
  "share_encoder_decoder_embeddings": true,
40
  "static_position_embeddings": true,
41
  "torch_dtype": "float16",
42
  "transformers_version": "4.34.0.dev0",
43
  "use_cache": true,
44
+ "vocab_size": 61247
45
  }
generation_config.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
  "bad_words_ids": [
3
  [
4
- 35697
5
  ]
6
  ],
7
  "bos_token_id": 0,
8
- "decoder_start_token_id": 35697,
9
- "eos_token_id": 26162,
10
- "forced_eos_token_id": 26162,
11
  "max_length": 512,
12
  "num_beams": 4,
13
- "pad_token_id": 35697,
14
  "renormalize_logits": true,
15
  "transformers_version": "4.34.0.dev0"
16
  }
 
1
  {
2
  "bad_words_ids": [
3
  [
4
+ 61246
5
  ]
6
  ],
7
  "bos_token_id": 0,
8
+ "decoder_start_token_id": 61246,
9
+ "eos_token_id": 25897,
10
+ "forced_eos_token_id": 25897,
11
  "max_length": 512,
12
  "num_beams": 4,
13
+ "pad_token_id": 61246,
14
  "renormalize_logits": true,
15
  "transformers_version": "4.34.0.dev0"
16
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:69ea2cf1906dd5768225f7ef1a778f8eaa87dbc8e2c1e4190ec2ca774975dd69
3
- size 425925172
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c4fb5fba22fc64947213c1adc0c9ac7f018cda71a466eea08231054eecf9851
3
+ size 478300630
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebbeee990e971cb41fd5262b168a044724e2c47dde06de7d10deb49287ad4cfc
3
- size 425982149
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6fc86f54d0ae5cea6773dfe7dd4474e2eddf331a4e21fadd02b6b76d4ca1fca
3
+ size 478357637