
    g+                         d Z ddlmZmZmZ er	 ddlmZ ddlmZ  ej	        e
          Z G d de          Z G d d	e          Zd
S )zMpt configuration    )TYPE_CHECKINGOptionalUnion   )PretrainedConfig)loggingc                   P     e Zd ZdZ	 	 	 	 	 	 	 	 	 	 d fd		Zedd            Z xZS )MptAttentionConfiga
  
    This is the configuration class to store the configuration of a [`MptAttention`] class. It is used to instantiate
    attention layers according to the specified arguments, defining the layers architecture. Instantiating a
    configuration with the defaults will yield a similar configuration to that of the MPT
    [mosaicml/mpt-7b](https://huggingface.co/mosaicml/mpt-7b) architecture. Most of the arguments are kept for backward
    compatibility with previous MPT models that are hosted on the Hub (previously with `trust_remote_code=True`).

    Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the
    documentation from [`PretrainedConfig`] for more information.

    Args:
        attn_type (`str`, *optional*, defaults to `"multihead_attention"`):
            type of attention to use. Options: `"multihead_attention"`, `"multiquery_attention"`.
        attn_pdrop (`float`, *optional*, defaults to 0.0):
            The dropout probability for the attention layers.
        attn_impl (`str`, *optional*, defaults to `"torch"`):
            The attention implementation to use. One of `"torch"`, `"flash"`, or `"triton"`.
        clip_qkv (`float`, *optional*):
            If not `None`, clip the queries, keys, and values in the attention layer to this value.
        softmax_scale (`float`, *optional*, defaults to `None`):
            If not `None`, scale the softmax in the attention layer by this value. If `None`, will default to
            `1/sqrt(hidden_size)`.
        prefix_lm (`bool`, *optional*, defaults to `False`)):
            Whether the model should operate as a Prefix LM. This requires passing an extra `prefix_mask` argument
            which indicates which tokens belong to the prefix. Tokens in the prefix can attend to one another
            bi-directionally. Tokens outside the prefix use causal attention.
        qk_ln (`bool`, *optional*, defaults to `False`):
            Whether to apply layer normalization to the queries and keys in the attention layer.
        attn_uses_sequence_id (`bool`, *optional*, defaults to `False`)):
            Whether to restrict attention to tokens that have the same token_type_ids. When the model is in `train`
            mode, this requires passing an extra *token_type_ids* argument which indicates which sub-sequence each
            token belongs to. Defaults to `False` meaning any provided *token_type_ids* will be ignored.
        alibi (`bool`, *optional*, defaults to `True`):
            Whether or not to use the alibi bias instead of positional embedding.
        alibi_bias_max (`int`, *optional*, defaults to 8):
            The maximum value of the alibi bias.
    multihead_attentionr   torchNFT   c                     t                                                       || _        || _        || _        || _        || _        || _        || _        |	| _	        || _
        |
| _        |dvrt          d|           d S )N)r   multiquery_attentionzX`attn_type` has to be either `multihead_attention` or `multiquery_attention`. Received: )super__init__	attn_type
attn_pdrop	attn_implclip_qkvsoftmax_scale	prefix_lmattn_uses_sequence_idalibiqk_lnalibi_bias_max
ValueError)selfr   r   r   r   r   r   r   r   r   r   kwargs	__class__s               e/var/www/html/ai-engine/env/lib/python3.11/site-packages/transformers/models/mpt/configuration_mpt.pyr   zMptAttentionConfig.__init__E   s     	"$" *"%:"

,KKKvktvv   LK    returnr   c                 N   |                      |            | j        |fi |\  }}|                    d          dk    r|d         }d|v rMt          | d          r=|d         | j        k    r,t
                              d|d          d| j         d            | j        |fi |S )N
model_typemptattn_configzYou are using a model of type z  to instantiate a model of type zN. This is not supported for all configurations of models and can yield errors.)_set_token_in_kwargsget_config_dictgethasattrr$   loggerwarning	from_dict)clspretrained_model_name_or_pathr   config_dicts       r    from_pretrainedz"MptAttentionConfig.from_pretrainedd   s      (((1c12OZZSYZZV??<((E11%m4K;&&73+E+E&+VbJcgjguJuJuNNr\1J r r>r r r  
 s}[33F333r!   )
r   r   r   NNFFFTr   )r"   r   )__name__
__module____qualname____doc__r   classmethodr1   __classcell__r   s   @r    r
   r
      s        $ $P (#     > 4 4 4 [4 4 4 4 4r!   r
   c            %            e Zd ZdZdZddddZ	 	 	 	 	 	 	 	 	 	 	 	 	 	 	 	 	 	 	 d&dedededededededededede	de
deeee
f                  d ed!ed"ed#e
d$ef$ fd%Z xZS )'	MptConfiga7  
    This is the configuration class to store the configuration of a [`MptModel`]. It is used to instantiate a Mpt model
    according to the specified arguments, defining the model architecture. Instantiating a configuration with the
    defaults will yield a similar configuration to the Mpt-7b architecture
    [mosaicml/mpt-7b](https://huggingface.co/mosaicml/mpt-7b).

    Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the
    documentation from [`PretrainedConfig`] for more information.


    Args:
        d_model (`int`, *optional*, defaults to 2048):
            Dimensionality of the embeddings and hidden states.
        n_heads (`int`, *optional*, defaults to 16):
            Number of attention heads for each attention layer in the Transformer encoder.
        n_layers (`int`, *optional*, defaults to 24):
            Number of hidden layers in the Transformer encoder.
        expansion_ratio (`int`, *optional*, defaults to 4):
            The ratio of the up/down scale in the MLP.
        max_seq_len (`int`, *optional*, defaults to 2048):
            The maximum sequence length of the model.
        vocab_size (`int`, *optional*, defaults to 50368):
            Vocabulary size of the Mpt model. Defines the maximum number of different tokens that can be represented by
            the `inputs_ids` passed when calling [`MptModel`]. Check [this
            discussion](https://huggingface.co/bigscience/mpt/discussions/120#633d28389addb8530b406c2a) on how the
            `vocab_size` has been defined.
        resid_pdrop (`float`, *optional*, defaults to 0.0):
            The dropout probability applied to the attention output before combining with residual.
        layer_norm_epsilon (`float`, *optional*, defaults to 1e-05):
            The epsilon to use in the layer normalization layers.
        emb_pdrop (`float`, *optional*, defaults to 0.0):
            The dropout probability for the embedding layer.
        learned_pos_emb (`bool`, *optional*, defaults to `True`):
            Whether to use learned positional embeddings.
        attn_config (`dict`, *optional*):
            A dictionary used to configure the model's attention module.
        init_device (`str`, *optional*, defaults to `"cpu"`):
            The device to use for parameter initialization. Defined for backward compatibility
        logit_scale (`float`, *optional*):
            If not None, scale the logits by this value.
        no_bias (`bool`, *optional*, defaults to `True`):
            Whether to use bias in all linear layers.
        verbose (`int`, *optional*, defaults to 0):
            The verbosity level to use for logging. Used in the previous versions of MPT models for logging. This
            argument is deprecated.
        embedding_fraction (`float`, *optional*, defaults to 1.0):
            The fraction to scale the gradients of the embedding layer by.
        norm_type (`str`, *optional*, defaults to `"low_precision_layernorm"`):
            Type of layer norm to use. All MPT models uses the same layer norm implementation. Defined for backward
            compatibility.
        use_cache (`bool`, *optional*, defaults to `False`):
            Whether or not the model should return the last key/values attentions (not used by all models).
        initializer_range (`float`, *optional*, defaults to 0.02):
            The standard deviation of the truncated_normal_initializer for initializing all weight matrices.

    Example:

    ```python
    >>> from transformers import MptConfig, MptModel

    >>> # Initializing a Mpt configuration
    >>> configuration = MptConfig()

    >>> # Initializing a model (with random weights) from the configuration
    >>> model = MptModel(configuration)

    >>> # Accessing the model configuration
    >>> configuration = model.config
    ```
    r%   n_headsd_modeln_layers)num_attention_headshidden_sizenum_hidden_layers                      h㈵>TNcpur         ?low_precision_layernormF{Gz?expansion_ratiomax_seq_len
vocab_sizeresid_pdroplayer_norm_epsilon	emb_pdroplearned_pos_embr&   init_devicelogit_scaleno_biasverboseembedding_fraction	norm_type	use_cachec                    |t                      | _        n.t          |t                    rt          di || _        n|| _        || _        || _        || _        || _        || _        || _	        || _
        |	| _        |
| _        || _        || _        || _        || _        || _        || _        || _        || _        || _         t-                      j        di | d S )N )r
   r&   
isinstancedictr<   r;   r=   rL   rM   rN   rO   rQ   rR   rS   rT   rU   rV   rW   rX   rP   rY   initializer_ranger   r   )r   r<   r;   r=   rL   rM   rN   rO   rP   rQ   rR   r&   rS   rT   rU   rV   rW   rX   rY   r^   r   r   s                        r    r   zMptConfig.__init__   s    . 133DT** 	+1@@K@@D*D .&$&".&&"4""4"!2""6"""""r!   )rA   rB   rC   rD   rA   rE   rF   rG   rF   TNrH   NTr   rI   rJ   FrK   )r2   r3   r4   r5   r$   attribute_mapintfloatboolr
   strr   r   r   r7   r8   s   @r    r:   r:   v   sf       E EN J( ' M   $( $*. 37$'2)/# /#/# /# 	/#
 /# /# /# /# "/# /# /# (/# /# eE3J/0/# /#  !/#" "#/#$ %/#& '/# /# /# /# /# /# /# /# /# /#r!   r:   N)r5   typingr   r   r   configuration_utilsr   utilsr   
get_loggerr2   r+   r
   r:   r[   r!   r    <module>rh      s      1 1 1 1 1 1 1 1 1 1  	 3 3 3 3 3 3       
	H	%	%U4 U4 U4 U4 U4) U4 U4 U4p~# ~# ~# ~# ~#  ~# ~# ~# ~# ~#r!   