
    Z j                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zMarian model configuration    )strict   )PreTrainedConfig)auto_docstringzHelsinki-NLP/opus-mt-en-de)
checkpointc                      ^  \ rS rSr% SrSrS/rSSSS.rS	r\	\
S
'   Sr\	S-  \
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\\	-  \
S'   Sr\\	-  \
S'   Sr\\
S'   Sr\\
S'   Sr\\
S'   Sr\	\
S'   Sr\\	-  \
S'   Sr\\	-  \
S '   Sr\\	-  \
S!'   S"r\\
S#'   S$r \	\
S%'   S&r!\\
S''   S$r"\	S-  \
S('   S)r#\	\$\	   -  S-  \
S*'   Sr%\	S-  \
S+'   S)r&\	\$\	   -  S-  \
S,'   Sr'\\
S-'   S&r(\\
S.'   Sr)\\
S/'   U 4S0 jr*S1r+U =r,$ )2MarianConfig   ah  
decoder_vocab_size (`int`, *optional*):
    Vocab size of the decoder layer's embedding.
share_encoder_decoder_embeddings (`bool`, *optional*, defaults to `True`):
    Whether to tie and share embeddings of encoder and decoder

Examples:

```python
>>> from transformers import MarianModel, MarianConfig

>>> # Initializing a Marian Helsinki-NLP/opus-mt-en-de style configuration
>>> configuration = MarianConfig()

>>> # Initializing a model from the Helsinki-NLP/opus-mt-en-de style configuration
>>> model = MarianModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```marianpast_key_valuesencoder_attention_headsd_modelencoder_layers)num_attention_headshidden_sizenum_hidden_layersi  
vocab_sizeNdecoder_vocab_sizei   max_position_embeddings   i   encoder_ffn_dim   decoder_layersdecoder_ffn_dimdecoder_attention_headsg        encoder_layerdropdecoder_layerdropT	use_cacheis_encoder_decodergeluactivation_functiong?dropoutattention_dropoutactivation_dropoutg{Gz?init_stdi  decoder_start_token_idFscale_embeddingpad_token_idr   eos_token_idbos_token_idforced_eos_token_id share_encoder_decoder_embeddings
is_decodertie_word_embeddingsc                 n   > U R                   =(       d    U R                  U l         [        TU ]  " S0 UD6  g )N )r   r   super__post_init__)selfkwargs	__class__s     ڀ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/marian/configuration_marian.pyr2   MarianConfig.__post_init__S   s*    "&"9"9"LT__''    )r   )-__name__
__module____qualname____firstlineno____doc__
model_typekeys_to_ignore_at_inferenceattribute_mapr   int__annotations__r   r   r   r   r   r   r   r   r   floatr   r   boolr   r!   strr   r"   r#   r$   r%   r&   r'   r(   r)   listr*   r+   r,   r-   r.   r2   __static_attributes____classcell__)r5   s   @r6   r	   r	      s   * J#4"58 -M J%)d
)#'S'NCOS#%S%NCOS#%S%%(us{(%(us{(It##%%GSGUS[%(us{(&))He"'C'!OT!$L#*$+,L#S	/D(,#L#*#23tCy4/3-1$d1J $$( (r8   r	   N)	r=   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r0   r8   r6   <module>rM      sH    ! . 3 # 78=(# =(  9=(@ 
r8   