Upload folder using huggingface_hub
Browse files- config.json +4 -0
- modeling_rwkv5.py +1 -1
config.json
CHANGED
|
@@ -2,6 +2,10 @@
|
|
| 2 |
"architectures": [
|
| 3 |
"Rwkv5ForCausalLM"
|
| 4 |
],
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
"attention_hidden_size": 2560,
|
| 6 |
"bos_token_id": 0,
|
| 7 |
"eos_token_id": 0,
|
|
|
|
| 2 |
"architectures": [
|
| 3 |
"Rwkv5ForCausalLM"
|
| 4 |
],
|
| 5 |
+
"auto_map": {
|
| 6 |
+
"AutoConfig": "configuration_rwkv5.Rwkv5Config",
|
| 7 |
+
"AutoModelForCausalLM": "modeling_rwkv5.Rwkv5ForCausalLM"
|
| 8 |
+
},
|
| 9 |
"attention_hidden_size": 2560,
|
| 10 |
"bos_token_id": 0,
|
| 11 |
"eos_token_id": 0,
|
modeling_rwkv5.py
CHANGED
|
@@ -436,7 +436,7 @@ class Rwkv5PreTrainedModel(PreTrainedModel):
|
|
| 436 |
"""
|
| 437 |
|
| 438 |
config_class = Rwkv5Config
|
| 439 |
-
base_model_prefix = "
|
| 440 |
_no_split_modules = ["RwkvBlock"]
|
| 441 |
_keep_in_fp32_modules = ["time_decay", "time_first"]
|
| 442 |
supports_gradient_checkpointing = True
|
|
|
|
| 436 |
"""
|
| 437 |
|
| 438 |
config_class = Rwkv5Config
|
| 439 |
+
base_model_prefix = "rwkv5"
|
| 440 |
_no_split_modules = ["RwkvBlock"]
|
| 441 |
_keep_in_fp32_modules = ["time_decay", "time_first"]
|
| 442 |
supports_gradient_checkpointing = True
|