
    cCi                     `    S r SSKJr  SSKJr  \R
                  " \5      r " S S\5      rS/r	g)z*Donut Swin Transformer model configuration   )PretrainedConfig)loggingc                   f   ^  \ rS rSrSrSrSSS.rSSS	S
/ SQ/ SQSSSSSSSSSS4U 4S jjrSrU =r	$ )DonutSwinConfig   a  
This is the configuration class to store the configuration of a [`DonutSwinModel`]. It is used to instantiate a
Donut model according to the specified arguments, defining the model architecture. Instantiating a configuration
with the defaults will yield a similar configuration to that of the Donut
[naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) architecture.

Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the
documentation from [`PretrainedConfig`] for more information.

Args:
    image_size (`int`, *optional*, defaults to 224):
        The size (resolution) of each image.
    patch_size (`int`, *optional*, defaults to 4):
        The size (resolution) of each patch.
    num_channels (`int`, *optional*, defaults to 3):
        The number of input channels.
    embed_dim (`int`, *optional*, defaults to 96):
        Dimensionality of patch embedding.
    depths (`list(int)`, *optional*, defaults to `[2, 2, 6, 2]`):
        Depth of each layer in the Transformer encoder.
    num_heads (`list(int)`, *optional*, defaults to `[3, 6, 12, 24]`):
        Number of attention heads in each layer of the Transformer encoder.
    window_size (`int`, *optional*, defaults to 7):
        Size of windows.
    mlp_ratio (`float`, *optional*, defaults to 4.0):
        Ratio of MLP hidden dimensionality to embedding dimensionality.
    qkv_bias (`bool`, *optional*, defaults to `True`):
        Whether or not a learnable bias should be added to the queries, keys and values.
    hidden_dropout_prob (`float`, *optional*, defaults to 0.0):
        The dropout probability for all fully connected layers in the embeddings and encoder.
    attention_probs_dropout_prob (`float`, *optional*, defaults to 0.0):
        The dropout ratio for the attention probabilities.
    drop_path_rate (`float`, *optional*, defaults to 0.1):
        Stochastic depth rate.
    hidden_act (`str` or `function`, *optional*, defaults to `"gelu"`):
        The non-linear activation function (function or string) in the encoder. If string, `"gelu"`, `"relu"`,
        `"selu"` and `"gelu_new"` are supported.
    use_absolute_embeddings (`bool`, *optional*, defaults to `False`):
        Whether or not to add absolute position embeddings to the patch embeddings.
    initializer_range (`float`, *optional*, defaults to 0.02):
        The standard deviation of the truncated_normal_initializer for initializing all weight matrices.
    layer_norm_eps (`float`, *optional*, defaults to 1e-05):
        The epsilon used by the layer normalization layers.

Example:

```python
>>> from transformers import DonutSwinConfig, DonutSwinModel

>>> # Initializing a Donut naver-clova-ix/donut-base style configuration
>>> configuration = DonutSwinConfig()

>>> # Randomly initializing a model from the naver-clova-ix/donut-base style configuration
>>> model = DonutSwinModel(configuration)

>>> # Accessing the model configuration
>>> configuration = model.config
```z
donut-swin	num_heads
num_layers)num_attention_headsnum_hidden_layers      r   `   )   r      r   )r   r      r      g      @Tg        g?geluFg{Gz?gh㈵>c                 L  > [         TU ]  " S0 UD6  Xl        X l        X0l        X@l        XPl        [        U5      U l        X`l	        Xpl
        Xl        Xl        Xl        Xl        Xl        Xl        Xl        UU l        Xl        [)        US[        U5      S-
  -  -  5      U l        g )Nr       )super__init__
image_size
patch_sizenum_channels	embed_dimdepthslenr	   r   window_size	mlp_ratioqkv_biashidden_dropout_probattention_probs_dropout_probdrop_path_rate
hidden_actuse_absolute_embeddingslayer_norm_epsinitializer_rangeinthidden_size)selfr   r   r   r   r   r   r   r    r!   r"   r#   r$   r%   r&   r(   r'   kwargs	__class__s                     l/home/james-whalen/.local/lib/python3.13/site-packages/transformers/models/donut/configuration_donut_swin.pyr   DonutSwinConfig.__init__[   s    ( 	"6"$$("f+"&" #6 ,H),$'>$,!2 y1Vq+AAB    )r#   r   r$   r   r%   r"   r*   r   r(   r'   r    r   r   r	   r   r!   r&   r   )
__name__
__module____qualname____firstlineno____doc__
model_typeattribute_mapr   __static_attributes____classcell__)r-   s   @r.   r   r      s]    9v J  +)M  %( %#)C )Cr0   r   N)
r5   configuration_utilsr   utilsr   
get_loggerr1   loggerr   __all__r   r0   r.   <module>r?      s>    1 3  
		H	%lC& lC^ 
r0   