
    Z jt
                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zALBERT model configuration    )strict   )PreTrainedConfig)auto_docstringzalbert/albert-xxlarge-v2)
checkpointc                   f   \ rS rSr% SrSrSr\\S'   Sr	\\S'   Sr
\\S	'   S
r\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\-  \S'   Sr\\-  \S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   S r\\-  \S!'   S"r\S#-  \S$'   Sr\S#-  \S%'   S&r\\\   -  S#-  \S''   S(r\\S)'   S*r g#)+AlbertConfig   aT  
num_hidden_groups (`int`, *optional*, defaults to 1):
    Number of groups for the hidden layers, parameters in the same group are shared.
inner_group_num (`int`, *optional*, defaults to 1):
    The number of inner repetition of attention and ffn.

Examples:

```python
>>> from transformers import AlbertConfig, AlbertModel

>>> # Initializing an ALBERT-xxlarge style configuration
>>> albert_xxlarge_configuration = AlbertConfig()

>>> # Initializing an ALBERT-base style configuration
>>> albert_base_configuration = AlbertConfig(
...     hidden_size=768,
...     num_attention_heads=12,
...     intermediate_size=3072,
... )

>>> # Initializing a model (with random weights) from the ALBERT-base style configuration
>>> model = AlbertModel(albert_xxlarge_configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```alberti0u  
vocab_size   embedding_sizei   hidden_size   num_hidden_layers   num_hidden_groups@   num_attention_headsi @  intermediate_sizeinner_group_numgelu_new
hidden_actg        hidden_dropout_probattention_probs_dropout_probi   max_position_embeddings   type_vocab_sizeg{Gz?initializer_rangeg-q=layer_norm_epsg?classifier_dropout_probr   Npad_token_idbos_token_idr   eos_token_idTtie_word_embeddings )!__name__
__module____qualname____firstlineno____doc__
model_typer   int__annotations__r   r   r   r   r   r   r   r   strr   floatr   r   r   r   r    r!   r"   r#   r$   listr%   bool__static_attributes__r&       ڀ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/albert/configuration_albert.pyr	   r	      s   8 JJNCKss!!"s"OS J '*u*03 #+3#&S&OS#u#!NE!+.S5[. L#*  L#* +,L#S	/D(, $$r4   r	   N)	r+   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r&   r4   r5   <module>r:      sH    ! . 3 # 562%# 2%  72%j 
r4   