
    Z j
                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zDistilBERT model configuration    )strict   )PreTrainedConfig)auto_docstringzgoogle/distilbert-base-uncased)
checkpointc                   N   \ rS rSr% SrSrSSSS.rSr\\	S	'   S
r
\\	S'   Sr\\	S'   Sr\\	S'   Sr\\	S'   Sr\\	S'   Sr\\	S'   Sr\\-  \	S'   Sr\\-  \	S'   Sr\\	S'   Sr\\	S'   Sr\\-  \	S'   Sr\\-  \	S'   Sr\S-  \	S'   Sr\\\   -  S-  \	S '   Sr\S-  \	S!'   S"r\\	S#'   S$rg)%DistilBertConfig   a  
sinusoidal_pos_embds (`boolean`, *optional*, defaults to `False`):
    Whether to use sinusoidal positional embeddings.
dim (`int`, *optional*, defaults to 768):
    Dimensionality of the encoder layers and the pooler layer.
qa_dropout (`float`, *optional*, defaults to 0.1):
    The dropout probabilities used in the question answering model [`DistilBertForQuestionAnswering`].
seq_classif_dropout (`float`, *optional*, defaults to 0.2):
    The dropout probabilities used in the sequence classification and the multiple choice model
    [`DistilBertForSequenceClassification`].

Examples:

```python
>>> from transformers import DistilBertConfig, DistilBertModel

>>> # Initializing a DistilBERT configuration
>>> configuration = DistilBertConfig()

>>> # Initializing a model (with random weights) from the configuration
>>> model = DistilBertModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```
distilbertdimn_headsn_layers)hidden_sizenum_attention_headsnum_hidden_layersi:w  
vocab_sizei   max_position_embeddingsFsinusoidal_pos_embds      i   i   
hidden_dimg?dropoutattention_dropoutgelu
activationg{Gz?initializer_range
qa_dropoutg?seq_classif_dropoutr   Npad_token_ideos_token_idbos_token_idTtie_word_embeddings )__name__
__module____qualname____firstlineno____doc__
model_typeattribute_mapr   int__annotations__r   r   boolr   r   r   r   r   floatr   r   strr   r   r   r   r    listr!   r"   __static_attributes__r#       ڈ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/distilbert/configuration_distilbert.pyr	   r	      s    4 J('M J#&S&!&$&HcGSCNJGUS[%(us{(J#u#!J!'** L#* +/L#S	/D(/#L#*# $$r2   r	   N)	r(   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r#   r2   r3   <module>r8      sH    % . 3 # ;<2%' 2%  =2%j 
r2   