
    Z j	                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zI-BERT configuration    )strict   )PreTrainedConfig)auto_docstringzkssteven/ibert-roberta-base)
checkpointc                   D   \ rS rSr% SrSrSr\\S'   Sr	\\S'   Sr
\\S	'   Sr\\S
'   Sr\\S'   Sr\\S'   Sr\\-  \S'   Sr\\-  \S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\S-  \S'   Sr\S-  \S'   Sr\\\   -  S-  \S'   S r\\S!'   S"r\\S#'   S$r\\S%'   S&rg)'IBertConfig   a  
type_vocab_size (`int`, *optional*, defaults to 2):
    The vocabulary size of the `token_type_ids` passed when calling [`IBertModel`]
quant_mode (`bool`, *optional*, defaults to `False`):
    Whether to quantize the model or not.
force_dequant (`str`, *optional*, defaults to `"none"`):
    Force dequantize specific nonlinear layer. Dequantized layers are then executed with full precision.
    `"none"`, `"gelu"`, `"softmax"`, `"layernorm"` and `"nonlinear"` are supported. As default, it is set as
    `"none"`, which does not dequantize any layers. Please specify `"gelu"`, `"softmax"`, or `"layernorm"` to
    dequantize GELU, Softmax, or LayerNorm, respectively. `"nonlinear"` will dequantize all nonlinear layers,
    i.e., GELU, Softmax, and LayerNorm.
iberti:w  
vocab_sizei   hidden_size   num_hidden_layersnum_attention_headsi   intermediate_sizegelu
hidden_actg?hidden_dropout_probattention_probs_dropout_probi   max_position_embeddings   type_vocab_sizeg{Gz?initializer_rangeg-q=layer_norm_eps   Npad_token_idr   bos_token_ideos_token_idF
quant_modenoneforce_dequantTtie_word_embeddings )__name__
__module____qualname____firstlineno____doc__
model_typer   int__annotations__r   r   r   r   r   strr   floatr   r   r   r   r   r   r   r   listr   boolr!   r"   __static_attributes__r#       ~/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/ibert/configuration_ibert.pyr	   r	      s     JJKs!!!s!J'**03 %#+3#&S&OS#u#!NE! L#*  L#* +,L#S	/D(,JM3 $$r1   r	   N)	r(   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r#   r1   r2   <module>r7      sG      . 3 # 89!%" !%  :!%H /r1   