
    cCiX                     R    S r SSKJr  \R                  " \5      r " S S5      rS/rg)zMMBT configuration   )loggingc                   "    \ rS rSrSrSS jrSrg)
MMBTConfig   a&  
This is the configuration class to store the configuration of a [`MMBTModel`]. It is used to instantiate a MMBT
model according to the specified arguments, defining the model architecture.

Args:
    config ([`PreTrainedConfig`]):
        Config of the underlying Transformer models. Its values are copied over to use a single config.
    num_labels (`int`, *optional*):
        Size of final Linear layer for classification.
    modal_hidden_size (`int`, *optional*, defaults to 2048):
        Embedding dimension of the non-text modality encoder.
Nc                 N    UR                   U l         X0l        U(       a  X l        g g )N__dict__modal_hidden_size
num_labels)selfconfigr   r
   s       p/home/james-whalen/.local/lib/python3.13/site-packages/transformers/models/deprecated/mmbt/configuration_mmbt.py__init__MMBTConfig.__init__&   s     !2(O     r   )Ni   )__name__
__module____qualname____firstlineno____doc__r   __static_attributes__ r   r   r   r      s    )r   r   N)r   utilsr   
get_loggerr   loggerr   __all__r   r   r   <module>r      s2       
		H	%) )* .r   