
    oi                     L    S SK JrJr  S SKJr  S SKJr  \ " S S\5      5       rg)    )	dataclassfield)PromptLearningConfig)PeftTypec                   f   ^  \ rS rSr% Sr\" SSS0S9r\\S'   \" SSS	0S9r	\
\S
'   U 4S jrSrU =r$ )PrefixTuningConfig   z
This is the configuration class to store the configuration of a [`PrefixEncoder`].

Args:
    encoder_hidden_size (`int`): The hidden size of the prompt encoder.
    prefix_projection (`bool`): Whether to project the prefix embeddings.
NhelpzThe hidden size of the encoder)defaultmetadataencoder_hidden_sizeFz$Whether to project the prefix tokensprefix_projectionc                 L   > [         TU ]  5         [        R                  U l        g )N)super__post_init__r   PREFIX_TUNING	peft_type)self	__class__s    Z/home/james-whalen/.local/lib/python3.13/site-packages/peft/tuners/prefix_tuning/config.pyr    PrefixTuningConfig.__post_init__(   s    !//    )r   )__name__
__module____qualname____firstlineno____doc__r   r   int__annotations__r   boolr   __static_attributes____classcell__)r   s   @r   r   r      sR      %:;   $@At 
0 0r   r   N)dataclassesr   r   peft.configr   
peft.utilsr   r    r   r   <module>r'      s,    ) ,  0- 0 0r   