
    Z j-
                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zPLBART model configuration    )strict   )PreTrainedConfig)auto_docstringzuclanlp/plbart-base)
checkpointc                      \ rS rSr% SrSrS/rSSSSS	.rS
r\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\\	-  \
S'   Sr\\	-  \
S'   Sr\\
S'   Sr\\
S'   Sr\\
S'   Sr\	\
S'   Sr\\	-  \
S'   Sr\\	-  \
S '   Sr\\	-  \
S!'   S"r\\
S'   Sr\\	-  \
S#'   Sr \\
S$'   S%r!\	S&-  \
S''   S(r"\	S&-  \
S)'   S*r#\	\$\	   -  S&-  \
S+'   S*r%\	\$\	   -  S&-  \
S,'   S-r&\\
S.'   Sr'\\
S/'   S0r(g&)1PLBartConfig   a  
Example:

```python
>>> from transformers import PLBartConfig, PLBartModel

>>> # Initializing a PLBART uclanlp/plbart-base style configuration
>>> configuration = PLBartConfig()

>>> # Initializing a model (with random weights) from the uclanlp/plbart-base style configuration
>>> model = PLBartModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```plbartpast_key_valuesencoder_attention_headsd_modelinit_stdencoder_layers)num_attention_headshidden_sizeinitializer_rangenum_hidden_layersiU  
vocab_sizei   max_position_embeddings   i   encoder_ffn_dim   decoder_layersdecoder_ffn_dimdecoder_attention_headsg        encoder_layerdropdecoder_layerdropT	use_cacheis_encoder_decodergeluactivation_functioni   g?dropoutattention_dropoutactivation_dropoutg{Gz?classifier_dropoutscale_embedding   Npad_token_idr   bos_token_id   eos_token_idforced_eos_token_idF
is_decodertie_word_embeddings ))__name__
__module____qualname____firstlineno____doc__
model_typekeys_to_ignore_at_inferenceattribute_mapr   int__annotations__r   r   r   r   r   r   r   r   floatr   r   boolr    r"   strr   r#   r$   r%   r   r&   r'   r)   r*   r,   listr-   r.   r/   __static_attributes__r0       ڀ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/plbart/configuration_plbart.pyr	   r	      sv     J#4"58 '-	M J#'S'NCOS#%S%NCOS#%S%%(us{(%(us{(It##%%GSGUS[%(us{(&))He&)) OT  L#*  L#* +,L#S	/D(,23tCy4/3J $$r@   r	   N)	r5   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r0   r@   rA   <module>rF      sH    ! . 3 # 013%# 3%  23%l 
r@   