Upload folder using huggingface_hub
Browse files- config.json +2 -6
- generation_config.json +1 -1
- model-00001-of-00003.safetensors +1 -1
- preprocessor_config.json +1 -1
config.json
CHANGED
|
@@ -10,7 +10,6 @@
|
|
| 10 |
"conf_conv_kernel_size": 5,
|
| 11 |
"conf_num_attention_heads": 8,
|
| 12 |
"conf_num_hidden_layers": 12,
|
| 13 |
-
"conf_positional_bias_size": 256,
|
| 14 |
"conf_reduction_factor": 4,
|
| 15 |
"conf_residual_weight": 0.5,
|
| 16 |
"gradient_clipping": 10000000000.0,
|
|
@@ -22,7 +21,6 @@
|
|
| 22 |
128,
|
| 23 |
32
|
| 24 |
],
|
| 25 |
-
"sscp_conv_eps": 0.001,
|
| 26 |
"sscp_conv_group_norm_eps": 0.001,
|
| 27 |
"sscp_conv_kernel_size": [
|
| 28 |
[
|
|
@@ -134,7 +132,6 @@
|
|
| 134 |
"altup_active_idx": 0,
|
| 135 |
"altup_coef_clip": 120.0,
|
| 136 |
"altup_correct_scale": true,
|
| 137 |
-
"altup_lr_multiplier": 1.0,
|
| 138 |
"altup_num_inputs": 4,
|
| 139 |
"attention_bias": false,
|
| 140 |
"attention_dropout": 0.0,
|
|
@@ -225,7 +222,6 @@
|
|
| 225 |
"num_hidden_layers": 35,
|
| 226 |
"num_key_value_heads": 2,
|
| 227 |
"num_kv_shared_layers": 15,
|
| 228 |
-
"query_pre_attn_scalar": 256,
|
| 229 |
"rms_norm_eps": 1e-06,
|
| 230 |
"rope_local_base_freq": 10000.0,
|
| 231 |
"rope_scaling": null,
|
|
@@ -237,11 +233,11 @@
|
|
| 237 |
"vocab_size_per_layer_input": 262144
|
| 238 |
},
|
| 239 |
"torch_dtype": "bfloat16",
|
| 240 |
-
"transformers_version": "4.53.
|
| 241 |
"unsloth_fixed": true,
|
| 242 |
"vision_config": {
|
| 243 |
"architecture": "mobilenetv5_300m_enc",
|
| 244 |
-
"do_pooling":
|
| 245 |
"hidden_size": 2048,
|
| 246 |
"initializer_range": 0.02,
|
| 247 |
"label_names": [
|
|
|
|
| 10 |
"conf_conv_kernel_size": 5,
|
| 11 |
"conf_num_attention_heads": 8,
|
| 12 |
"conf_num_hidden_layers": 12,
|
|
|
|
| 13 |
"conf_reduction_factor": 4,
|
| 14 |
"conf_residual_weight": 0.5,
|
| 15 |
"gradient_clipping": 10000000000.0,
|
|
|
|
| 21 |
128,
|
| 22 |
32
|
| 23 |
],
|
|
|
|
| 24 |
"sscp_conv_group_norm_eps": 0.001,
|
| 25 |
"sscp_conv_kernel_size": [
|
| 26 |
[
|
|
|
|
| 132 |
"altup_active_idx": 0,
|
| 133 |
"altup_coef_clip": 120.0,
|
| 134 |
"altup_correct_scale": true,
|
|
|
|
| 135 |
"altup_num_inputs": 4,
|
| 136 |
"attention_bias": false,
|
| 137 |
"attention_dropout": 0.0,
|
|
|
|
| 222 |
"num_hidden_layers": 35,
|
| 223 |
"num_key_value_heads": 2,
|
| 224 |
"num_kv_shared_layers": 15,
|
|
|
|
| 225 |
"rms_norm_eps": 1e-06,
|
| 226 |
"rope_local_base_freq": 10000.0,
|
| 227 |
"rope_scaling": null,
|
|
|
|
| 233 |
"vocab_size_per_layer_input": 262144
|
| 234 |
},
|
| 235 |
"torch_dtype": "bfloat16",
|
| 236 |
+
"transformers_version": "4.53.1",
|
| 237 |
"unsloth_fixed": true,
|
| 238 |
"vision_config": {
|
| 239 |
"architecture": "mobilenetv5_300m_enc",
|
| 240 |
+
"do_pooling": false,
|
| 241 |
"hidden_size": 2048,
|
| 242 |
"initializer_range": 0.02,
|
| 243 |
"label_names": [
|
generation_config.json
CHANGED
|
@@ -9,5 +9,5 @@
|
|
| 9 |
"pad_token_id": 0,
|
| 10 |
"top_k": 64,
|
| 11 |
"top_p": 0.95,
|
| 12 |
-
"transformers_version": "4.53.
|
| 13 |
}
|
|
|
|
| 9 |
"pad_token_id": 0,
|
| 10 |
"top_k": 64,
|
| 11 |
"top_p": 0.95,
|
| 12 |
+
"transformers_version": "4.53.1"
|
| 13 |
}
|
model-00001-of-00003.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 3723417366
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:e6d2fa805364c1245744fa9efc4a8aaf17cd2c866ed385562f85bd50de684393
|
| 3 |
size 3723417366
|
preprocessor_config.json
CHANGED
|
@@ -41,7 +41,7 @@
|
|
| 41 |
"processor_class": "Gemma3nProcessor",
|
| 42 |
"resample": 2,
|
| 43 |
"rescale_factor": 0.00392156862745098,
|
| 44 |
-
"return_attention_mask":
|
| 45 |
"return_tensors": null,
|
| 46 |
"sampling_rate": 16000,
|
| 47 |
"size": {
|
|
|
|
| 41 |
"processor_class": "Gemma3nProcessor",
|
| 42 |
"resample": 2,
|
| 43 |
"rescale_factor": 0.00392156862745098,
|
| 44 |
+
"return_attention_mask": true,
|
| 45 |
"return_tensors": null,
|
| 46 |
"sampling_rate": 16000,
|
| 47 |
"size": {
|