
    Z j                     p    S SK Jr  SSKJr  SSKJr  SSKJrJr  \" SS9\ " S	 S
\5      5       5       r	S
/r
g)    )strict   )PreTrainedConfig)auto_docstring   )CONFIG_MAPPING
AutoConfigznvidia/music-flamingo-2601-hf)
checkpointc                      ^  \ rS rSr% SrSr\\S.rSr\	\
-  S-  \S'   Sr\	\
-  S-  \S'   Sr\\S	'   S
r\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\\S'   Sr\	S-  \S'   U 4S jrSrU =r$ )MusicFlamingoConfig   a5  
audio_bos_token_id (`int`, *optional*, defaults to 151670):
    The beginning-of-audio token index used to mark the start of audio spans.
audio_eos_token_id (`int`, *optional*, defaults to 151671):
    The end-of-audio token index used to mark the end of audio spans.
audio_frame_step (`float`, *optional*, defaults to 0.01):
    Duration in seconds of one input mel frame (trained with hop_length 160 at sampling_rate 16000).

Example:

```python
>>> from transformers import MusicFlamingoForConditionalGeneration, MusicFlamingoConfig, AudioFlamingo3EncoderConfig, Qwen2Config

>>> # Initializing an MusicFlamingoEncoder config
>>> audio_config = AudioFlamingo3EncoderConfig()

>>> # Initializing a Qwen2 config
>>> text_config = Qwen2Config()

>>> # Initializing an MusicFlamingo configuration
>>> configuration = MusicFlamingoConfig(audio_config, text_config)

>>> # Initializing a model from the musicflamingo style configuration
>>> model = MusicFlamingoForConditionalGeneration(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```musicflamingo)audio_configtext_configNr   r   iuP audio_token_idgeluprojector_hidden_actTprojector_biasivP audio_bos_token_idiwP audio_eos_token_idg{Gz?audio_frame_steprope_parametersc                   > [        U R                  [        5      (       aN  U R                  S   S;   a  SU R                  S'   [        U R                  S      " S
0 U R                  D6U l        O U R                  c  [        S   " 5       U l        [        U R                  [        5      (       aU  U R                  R                  SS5      U R                  S'   [        U R                  S      " S
0 U R                  D6U l        O U R                  c  [        S   " 5       U l        U R                  c  SSSS.U l        U R                  S	   U l        U R                  R                  U l	        [        TU ],  " S
0 UD6  g )N
model_type)Nmusicflamingo_encoderaudioflamingo3_encoderqwen2defaulti  g?)	rope_type
rope_thetapartial_rotary_factorr     )
isinstancer   dictr   r   getr   max_position_embeddingshidden_sizehead_dimsuper__post_init__)selfkwargs	__class__s     ڎ/root/GenerationalWealth/GenerationalWealth/venv/lib/python3.13/site-packages/transformers/models/musicflamingo/configuration_musicflamingo.pyr*   !MusicFlamingoConfig.__post_init__K   sJ   d''..  .2QQ2J!!,/ .t/@/@/N O dRVRcRc dD& ./G H JDd&&---1-=-=-A-A,PW-XD\*-d.>.>|.LMaPTP`P`aD%-g68D'1:$il#mD '+';';L'I$))55''    )r   r(   r&   r   r   )__name__
__module____qualname____firstlineno____doc__r   r	   sub_configsr   r$   r   __annotations__r   r   intr   strr   boolr   r   r   floatr   r*   __static_attributes____classcell__)r-   s   @r.   r   r      s    : !J#-jIK37L$))D0726K((4/6 NC  &#&ND$$$$"e"#'OTD['( (r0   r   N)huggingface_hub.dataclassesr   configuration_utilsr   utilsr   autor   r	   r   __all__r"   r0   r.   <module>rC      sH   . / 3 # - :;>(* >(  <>(B !
!r0   