
    Z jZ
                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zCodeGen model configuration    )strict   )PreTrainedConfig)auto_docstringzSalesforce/codegen-2B-mono)
checkpointc                   X   \ rS rSr% SrSrSSSSS.rS	r\\	S
'   Sr
\\	S'   Sr\\	S'   Sr\\	S'   Sr\\	S'   Sr\\	S'   Sr\\	S'   Sr\S-  \	S'   Sr\\	S'   Sr\\-  \	S'   Sr\\-  \	S'   Sr\\-  \	S'   Sr\\	S'   Sr\\	S'   Sr\\	S'   S r\S-  \	S!'   S r\\\   -  S-  \	S"'   S#r\\	S$'   S%rg)&CodeGenConfig   a>  
n_ctx (`int`, *optional*, defaults to 2048):
    This attribute is used in `CodeGenModel.__init__` without any real effect.
    The dropout probability for all fully connected layers in the embeddings, encoder, and pooler.
rotary_dim (`int`, *optional*, defaults to 64):
    Number of dimensions in the embedding that Rotary Position Embedding is applied to.
n_inner (`int`, *optional*):
    Dimensionality of the inner feed-forward layers. `None` will set it to 4 times n_embd

Example:

```python
>>> from transformers import CodeGenConfig, CodeGenModel

>>> # Initializing a CodeGen 6B configuration
>>> configuration = CodeGenConfig()

>>> # Initializing a model (with random weights) from the configuration
>>> model = CodeGenModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```codegenn_positionsn_embdn_headn_layer)max_position_embeddingshidden_sizenum_attention_headsnum_hidden_layersi  
vocab_sizei   n_ctxi         @   
rotary_dimNn_innergelu_newactivation_functiong        resid_pdrop
embd_pdrop
attn_pdropgh㈵>layer_norm_epsilong{Gz?initializer_rangeT	use_cacheiP  bos_token_ideos_token_idFtie_word_embeddings ) __name__
__module____qualname____firstlineno____doc__
model_typeattribute_mapr   int__annotations__r   r   r   r   r   r   r   r   strr   floatr   r   r    r!   r"   boolr#   r$   listr%   __static_attributes__r&       ڂ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/codegen/configuration_codegen.pyr	   r	      s    0 J#0'&	M JKE3FCGSFCJGS4Z))"K"!J!!J! $$#u#It$L#*$+0L#S	/D(0 %%r5   r	   N)	r+   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r&   r5   r6   <module>r;      sH    " . 3 # 782&$ 2&  92&j 
r5   