
    Z j{	                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zConvBERT model configuration    )strict   )PreTrainedConfig)auto_docstringzYituTech/conv-bert-base)
checkpointc                      \ rS rSr% SrSrSr\\S'   Sr	\\S'   Sr
\\S	'   Sr\\S
'   Sr\\S'   Sr\\S'   Sr\\-  \S'   Sr\\-  \S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\S-  \S'   Sr\S-  \S'   Sr\\\   -  S-  \S'   Sr\\S '   Sr\\S!'   S"r\\S#'   Sr\\S$'   Sr\\-  S-  \S%'   S&r\ \S''   S&r!\ \S('   S)r"\ \S*'   S+r#g),ConvBertConfig   aw  
head_ratio (`int`, *optional*, defaults to 2):
    Ratio gamma to reduce the number of attention heads.
num_groups (`int`, *optional*, defaults to 1):
    The number of groups for grouped linear layers for ConvBert model

Example:

```python
>>> from transformers import ConvBertConfig, ConvBertModel

>>> # Initializing a ConvBERT convbert-base-uncased style configuration
>>> configuration = ConvBertConfig()

>>> # Initializing a model (with random weights) from the convbert-base-uncased style configuration
>>> model = ConvBertModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```convberti:w  
vocab_sizei   hidden_size   num_hidden_layersnum_attention_headsi   intermediate_sizegelu
hidden_actg?hidden_dropout_probattention_probs_dropout_probi   max_position_embeddings   type_vocab_sizeg{Gz?initializer_rangeg-q=layer_norm_eps   Npad_token_idr   bos_token_ideos_token_idembedding_size
head_ratio	   conv_kernel_size
num_groupsclassifier_dropoutF
is_decoderadd_cross_attentionTtie_word_embeddings )$__name__
__module____qualname____firstlineno____doc__
model_typer   int__annotations__r   r   r   r   r   strr   floatr   r   r   r   r   r   r   r   listr   r    r"   r#   r$   r%   boolr&   r'   __static_attributes__r(       ڄ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/convbert/configuration_convbert.pyr	   r	      s%   * JJKs!!!s!J'**03 %#+3#&S&OS#u#!NE! L#*  L#* +,L#S	/D(,NCJcJ-1d*1J %% $$r6   r	   N)	r-   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r(   r6   r7   <module>r<      sH    # . 3 # 45.%% .%  6.%b 
r6   