
    Z j
                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zRWKV configuration    )strict   )PreTrainedConfig)auto_docstringzRWKV/rwkv-4-169m-pile)
checkpointc                     ^  \ rS rSr% SrSrSS0rSr\\	S'   Sr
\\	S'   S	r\\	S
'   Sr\\	S'   Sr\S-  \	S'   Sr\S-  \	S'   Sr\\	S'   Sr\S-  \	S'   Sr\\\   -  S-  \	S'   Sr\\	S'   Sr\\	S'   Sr\\	S'   U 4S jrSrU =r$ )
RwkvConfig   a  
context_length (`int`, *optional*, defaults to 1024):
    The maximum sequence length that this model can be used with in a single forward (using it in RNN mode
    lets use any sequence length).
attention_hidden_size (`int`, *optional*):
    Dimensionality of the attention hidden states. Will default to `hidden_size` if unset.
rescale_every (`int`, *optional*, defaults to 6):
    At inference, the hidden states (and weights of the corresponding output layers) are divided by 2 every
    `rescale_every` layer. If set to 0 or a negative number, no rescale is done.

Example:

```python
>>> from transformers import RwkvConfig, RwkvModel

>>> # Initializing a Rwkv configuration
>>> configuration = RwkvConfig()

>>> # Initializing a model (with random weights) from the configuration
>>> model = RwkvModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```rwkvmax_position_embeddingscontext_lengthie  
vocab_sizei   i   hidden_size    num_hidden_layersNattention_hidden_sizeintermediate_sizegh㈵>layer_norm_epsilonr   bos_token_ideos_token_id   rescale_everyFtie_word_embeddingsT	use_cachec                    > U R                   b  U R                   OU R                  U l         U R                  b  U R                  OSU R                  -  U l        [        TU ]  " S0 UD6  g )N    )r   r   r   super__post_init__)selfkwargs	__class__s     |/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/rwkv/configuration_rwkv.pyr   RwkvConfig.__post_init__C   sc    *.*D*D*PD&&VZVfVf 	" <@;Q;Q;]!7!7cdgkgwgwcw''    )r   r   )__name__
__module____qualname____firstlineno____doc__
model_typeattribute_mapr   int__annotations__r   r   r   r   r   r   floatr   r   listr   r   boolr   r   __static_attributes____classcell__)r"   s   @r#   r	   r	      s    2 J.0@AMJNCKs(,3:,$(sTz( $$ L#* +,L#S	/D(,M3 %%It( (r%   r	   N)	r*   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r   r%   r#   <module>r8      sG     . 3 # 230(! 0(  40(f .r%   