Commit
·
214ede7
1
Parent(s):
9d2f75e
Fix eos_token_id to align with vocabulary of this model
Browse filesSame problem as the 40b model in this repo.
See https://huggingface.co/TheBloke/falcon-40b-instruct-GPTQ/discussions/8
Upstream discussion: https://huggingface.co/tiiuae/falcon-40b-instruct/discussions/11
- configuration_RW.py +1 -1
configuration_RW.py
CHANGED
|
@@ -38,7 +38,7 @@ class RWConfig(PretrainedConfig):
|
|
| 38 |
initializer_range=0.02,
|
| 39 |
use_cache=True,
|
| 40 |
bos_token_id=1,
|
| 41 |
-
eos_token_id=
|
| 42 |
apply_residual_connection_post_layernorm=False,
|
| 43 |
hidden_dropout=0.0,
|
| 44 |
attention_dropout=0.0,
|
|
|
|
| 38 |
initializer_range=0.02,
|
| 39 |
use_cache=True,
|
| 40 |
bos_token_id=1,
|
| 41 |
+
eos_token_id=11,
|
| 42 |
apply_residual_connection_post_layernorm=False,
|
| 43 |
hidden_dropout=0.0,
|
| 44 |
attention_dropout=0.0,
|