
    Z j	                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zPEGASUS model configuration    )strict   )PreTrainedConfig)auto_docstringzgoogle/pegasus-large)
checkpointc                      \ rS rSr% SrSrS/rSSSS.rS	r\	\
S
'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\	\
S'   Sr\\	-  \
S'   Sr\\	-  \
S'   Sr\\
S'   Sr\\
S'   Sr\\
S'   Sr\	\
S'   Sr\\	-  \
S'   Sr\\	-  \
S'   Sr\\	-  \
S'   S r\\
S!'   S"r\	S#-  \
S$'   S%r \\
S&'   S"r!\	S#-  \
S''   S(r"\	\#\	   -  S#-  \
S)'   S(r$\	\#\	   -  S#-  \
S*'   S%r%\\
S+'   Sr&\\
S,'   S-r'g#).PegasusConfig   a  
Example:

```python
>>> from transformers import PegasusConfig, PegasusModel

>>> # Initializing a PEGASUS google/pegasus-large style configuration
>>> configuration = PegasusConfig()

>>> # Initializing a model (with random weights) from the google/pegasus-large style configuration
>>> model = PegasusModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```pegasuspast_key_valuesencoder_attention_headsd_modelencoder_layers)num_attention_headshidden_sizenum_hidden_layersiY  
vocab_sizei   max_position_embeddings   i   encoder_ffn_dim   decoder_layersdecoder_ffn_dimdecoder_attention_headsg        encoder_layerdropdecoder_layerdropT	use_cacheis_encoder_decodergeluactivation_functiong?dropoutattention_dropoutactivation_dropoutg{Gz?init_stdr   Ndecoder_start_token_idFscale_embeddingpad_token_id   eos_token_idforced_eos_token_id
is_decodertie_word_embeddings )(__name__
__module____qualname____firstlineno____doc__
model_typekeys_to_ignore_at_inferenceattribute_mapr   int__annotations__r   r   r   r   r   r   r   r   floatr   r   boolr   r    strr   r!   r"   r#   r$   r%   r&   r'   r)   listr*   r+   r,   __static_attributes__r-       ڂ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/pegasus/configuration_pegasus.pyr	   r	      se     J#4"58 -M J#'S'NCOS#%S%NCOS#%S%%(us{(%(us{(It##%%GSGUS[%(us{(&))He)*C$J*!OT! L#* +,L#S	/D(,23tCy4/3J $$r=   r	   N)	r2   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r-   r=   r>   <module>rC      sH    " . 3 # 121%$ 1%  31%h 
r=   