
    Z j                         S SK Jr  SSKJr  SSKJr  \" SS9\ " S S\5      5       5       r\" SS9\ " S	 S
\5      5       5       rS
/rg)    )strict   )PreTrainedConfig)auto_docstringz&PaddlePaddle/SLANeXt_wired_safetensors)
checkpointc                   D   \ rS rSr% SrSrSr\\S'   Sr	\\S'   Sr
\\S	'   Sr\\S
'   Sr\\S'   Sr\\S'   Sr\\\   -  \\\4   -  \S'   Sr\\S'   Sr\\S'   Sr\\-  \S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\   \\S 4   -  \S!'   S"r\\S#'   S$rg%)&SLANeXtVisionConfig   a  
output_channels (`int`, *optional*, defaults to 256):
    Dimensionality of the output channels in the Patch Encoder.
use_abs_pos (`bool`, *optional*, defaults to `True`):
    Whether to use absolute position embedding.
use_rel_pos (`bool`, *optional*, defaults to `True`):
    Whether to use relative position embedding.
window_size (`int`, *optional*, defaults to 14):
    Window size for relative position.
global_attn_indexes (`list[int]`, *optional*, defaults to `[2, 5, 8, 11]`):
    The indexes of the global attention layers.
mlp_dim (`int`, *optional*, defaults to 3072):
    The dimensionality of the MLP layer in the Transformer encoder.
vision_configi   hidden_size   output_channels   num_hidden_layersnum_attention_headsr   num_channels   
image_size   
patch_sizegelu
hidden_actgư>layer_norm_epsg        attention_dropoutg|=initializer_rangeTqkv_biasuse_abs_posuse_rel_pos   window_size)            .global_attn_indexesi   mlp_dim N)__name__
__module____qualname____firstlineno____doc__base_config_keyr   int__annotations__r   r   r   r   r   r   listtupler   strr   floatr   r   r   boolr   r   r    r%   r&   __static_attributes__r'       ڂ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/slanext/configuration_slanext.pyr	   r	      s     &OKOSs!!L#J46Jd3i%S/16J!NE!%(us{($u$HdKKK7DcU38_4DGSr6   r	   c                      ^  \ rS rSr% SrSrS\0rSr\	\-  S-  \
S'   Sr\\
S'   Sr\\
S	'   S
r\\
S'   Sr\\
S'   Sr\\
S'   U 4S jrSrU =r$ )SLANeXtConfigB   a  
vision_config (`dict` or [`SLANeXtVisionConfig`], *optional*):
    Configuration for the vision encoder. If `None`, a default [`SLANeXtVisionConfig`] is used.
post_conv_in_channels (`int`, *optional*, defaults to 256):
    Number of input channels for the post-encoder convolution layer.
post_conv_out_channels (`int`, *optional*, defaults to 512):
    Number of output channels for the post-encoder convolution layer.
out_channels (`int`, *optional*, defaults to 50):
    Vocabulary size for the table structure token prediction head, i.e., the number of distinct structure
    tokens the model can predict.
hidden_size (`int`, *optional*, defaults to 512):
    Dimensionality of the hidden states in the attention GRU cell and the structure/location prediction heads.
max_text_length (`int`, *optional*, defaults to 500):
    Maximum number of autoregressive decoding steps (tokens) for the structure and location decoder.
slanextr   Nr   post_conv_in_channelsr   post_conv_out_channels2   out_channelsr   i  max_text_lengthc                    > U R                   c  [        5       U l         O9[        U R                   [        5      (       a  [        S0 U R                   D6U l         [        TU ]  " S0 UD6  g )Nr'   )r   r	   
isinstancedictsuper__post_init__)selfkwargs	__class__s     r7   rE   SLANeXtConfig.__post_init___   sS    %!4!6D**D11!4!Jt7I7I!JD''r6   )r   )r(   r)   r*   r+   r,   
model_typer	   sub_configsr   rC   r/   r<   r.   r=   r?   r   r@   rE   r5   __classcell__)rH   s   @r7   r9   r9   B   sm      J"$78K7;M4--4;!$3$"%C%L#KOS( (r6   r9   N)	huggingface_hub.dataclassesr   configuration_utilsr   utilsr   r	   r9   __all__r'   r6   r7   <module>rQ      sr   , / 3 # CD!* !  E!H CD ($  (  E (F 
r6   