seba commited on
Commit
3077a55
·
verified ·
1 Parent(s): 9ebc396

Upload folder using huggingface_hub

Browse files
fsq_layer.mlmodelc/analytics/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c681f1760e3627e783d507ac461c12ba200de1eb8f2684006d04733594749b3
3
+ size 243
fsq_layer.mlmodelc/coremldata.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91a8c6ee76db8383776831e6d600571905d5642461ab3b9ba51fc0a25b36fa58
3
+ size 207
fsq_layer.mlmodelc/model.mil ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ program(1.3)
2
+ [buildInfo = dict<string, string>({{"coremlc-component-MIL", "3405.2.1"}, {"coremlc-version", "3405.2.1"}})]
3
+ {
4
+ func main<ios18>(tensor<fp16, [1, 1024, 1, ?]> x) [FlexibleShapeInformation = tuple<tuple<string, dict<string, tensor<int32, [?]>>>, tuple<string, dict<string, dict<string, tensor<int32, [?]>>>>>((("DefaultShapes", {{"x", [1, 1024, 1, 32]}}), ("EnumeratedShapes", {{"2a8dd4d2", {{"x", [1, 1024, 1, 64]}}}, {"3bf9b6d7", {{"x", [1, 1024, 1, 128]}}}, {"44109b23", {{"x", [1, 1024, 1, 1]}}}, {"759fca31", {{"x", [1, 1024, 1, 8]}}}, {"d30434c5", {{"x", [1, 1024, 1, 16]}}}, {"e483fcaa", {{"x", [1, 1024, 1, 4]}}}, {"fabb4dad", {{"x", [1, 1024, 1, 32]}}}})))] {
5
+ string hidden_3_pad_type_0 = const()[name = string("hidden_3_pad_type_0"), val = string("valid")];
6
+ tensor<int32, [2]> hidden_3_strides_0 = const()[name = string("hidden_3_strides_0"), val = tensor<int32, [2]>([1, 1])];
7
+ tensor<int32, [4]> hidden_3_pad_0 = const()[name = string("hidden_3_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
8
+ tensor<int32, [2]> hidden_3_dilations_0 = const()[name = string("hidden_3_dilations_0"), val = tensor<int32, [2]>([1, 1])];
9
+ int32 hidden_3_groups_0 = const()[name = string("hidden_3_groups_0"), val = int32(1)];
10
+ tensor<fp16, [256, 1024, 1, 1]> var_8_to_fp16 = const()[name = string("op_8_to_fp16"), val = tensor<fp16, [256, 1024, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(64)))];
11
+ tensor<fp16, [256]> in_proj_bias_to_fp16 = const()[name = string("in_proj_bias_to_fp16"), val = tensor<fp16, [256]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(524416)))];
12
+ tensor<fp16, [1, 256, 1, ?]> hidden_3_cast_fp16 = conv(bias = in_proj_bias_to_fp16, dilations = hidden_3_dilations_0, groups = hidden_3_groups_0, pad = hidden_3_pad_0, pad_type = hidden_3_pad_type_0, strides = hidden_3_strides_0, weight = var_8_to_fp16, x = x)[name = string("hidden_3_cast_fp16")];
13
+ tensor<fp16, [1, 256, 1, ?]> hidden_cast_fp16 = tanh(x = hidden_3_cast_fp16)[name = string("hidden_cast_fp16")];
14
+ fp16 var_16_promoted_to_fp16 = const()[name = string("op_16_promoted_to_fp16"), val = fp16(0x1.2p+3)];
15
+ tensor<fp16, [1, 256, 1, ?]> var_17_cast_fp16 = mul(x = hidden_cast_fp16, y = var_16_promoted_to_fp16)[name = string("op_17_cast_fp16")];
16
+ tensor<fp16, [1, 256, 1, ?]> var_18_cast_fp16 = round(x = var_17_cast_fp16)[name = string("op_18_cast_fp16")];
17
+ fp16 _inversed_x_y_0_to_fp16 = const()[name = string("_inversed_x_y_0_to_fp16"), val = fp16(0x1.c7p-4)];
18
+ tensor<fp16, [1, 256, 1, ?]> _inversed_x_cast_fp16 = mul(x = var_18_cast_fp16, y = _inversed_x_y_0_to_fp16)[name = string("_inversed_x_cast_fp16")];
19
+ string var_31_pad_type_0 = const()[name = string("op_31_pad_type_0"), val = string("valid")];
20
+ tensor<int32, [2]> var_31_strides_0 = const()[name = string("op_31_strides_0"), val = tensor<int32, [2]>([1, 1])];
21
+ tensor<int32, [4]> var_31_pad_0 = const()[name = string("op_31_pad_0"), val = tensor<int32, [4]>([0, 0, 0, 0])];
22
+ tensor<int32, [2]> var_31_dilations_0 = const()[name = string("op_31_dilations_0"), val = tensor<int32, [2]>([1, 1])];
23
+ int32 var_31_groups_0 = const()[name = string("op_31_groups_0"), val = int32(1)];
24
+ tensor<fp16, [1024, 256, 1, 1]> var_25_to_fp16 = const()[name = string("op_25_to_fp16"), val = tensor<fp16, [1024, 256, 1, 1]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(524992)))];
25
+ tensor<fp16, [1024]> out_proj_bias_to_fp16 = const()[name = string("out_proj_bias_to_fp16"), val = tensor<fp16, [1024]>(BLOBFILE(path = string("@model_path/weights/weight.bin"), offset = uint64(1049344)))];
26
+ tensor<fp16, [1, 1024, 1, ?]> output = conv(bias = out_proj_bias_to_fp16, dilations = var_31_dilations_0, groups = var_31_groups_0, pad = var_31_pad_0, pad_type = var_31_pad_type_0, strides = var_31_strides_0, weight = var_25_to_fp16, x = _inversed_x_cast_fp16)[name = string("op_31_cast_fp16")];
27
+ } -> (output);
28
+ }
fsq_layer.mlmodelc/weights/weight.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff2cd9a8fb4996f3ff2905b7fb86a439a176f0a2b10cc527523d5385f8e43291
3
+ size 1051456