
    Z j
                     d    S r SSKJr  SSKJr  SSKJr  \" SS9\ " S S	\5      5       5       rS	/rg
)zLUKE configuration    )strict   )PreTrainedConfig)auto_docstringzstudio-ousia/luke-base)
checkpointc                   l   \ rS rSr% SrSrSr\\S'   Sr	\\S'   Sr
\\S	'   S
r\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\-  \S'   Sr\\-  \S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   S r\\-  S -  \S!'   S"r\S -  \S#'   S$r\S -  \S%'   Sr\\\   -  S -  \S&'   Sr\\S''   S(r g ))
LukeConfig   a  
entity_vocab_size (`int`, *optional*, defaults to 500000):
    Entity vocabulary size of the LUKE model. Defines the number of different entities that can be represented
    by the `entity_ids` passed when calling [`LukeModel`].
entity_emb_size (`int`, *optional*, defaults to 256):
    The number of dimensions of the entity embedding.
use_entity_aware_attention (`bool`, *optional*, defaults to `True`):
    Whether or not the model should use the entity-aware self-attention mechanism proposed in [LUKE: Deep
    Contextualized Entity Representations with Entity-aware Self-attention (Yamada et
    al.)](https://huggingface.co/papers/2010.01057).

Examples:

```python
>>> from transformers import LukeConfig, LukeModel

>>> # Initializing a LUKE configuration
>>> configuration = LukeConfig()

>>> # Initializing a model from the configuration
>>> model = LukeModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```lukei[  
vocab_sizei  entity_vocab_sizei   hidden_size   entity_emb_size   num_hidden_layersnum_attention_headsi   intermediate_sizegelu
hidden_actg?hidden_dropout_probattention_probs_dropout_probi   max_position_embeddings   type_vocab_sizeg{Gz?initializer_rangeg-q=layer_norm_epsTuse_entity_aware_attentionNclassifier_dropout   pad_token_idr   bos_token_ideos_token_idtie_word_embeddings )!__name__
__module____qualname____firstlineno____doc__
model_typer   int__annotations__r   r   r   r   r   r   r   strr   floatr   r   r   r   r   r   boolr   r!   r"   r#   listr$   __static_attributes__r%       |/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/luke/configuration_luke.pyr	   r	      s   4 JJ#s#KOSs!!!s!J'**03 %#+3#&S&OS#u#!NE!'++-1d*1 L#*  L#* +,L#S	/D(, $$r3   r	   N)	r*   huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   __all__r%   r3   r4   <module>r9      sG     . 3 # 340%! 0%  50%f .r3   