
    <ho(                         S r SSKJrJr  SSKrSSKJr  SSKJ	r	  SSK
JrJrJrJrJr  SSKJrJr   " S	 S
\SS9r " S S\SS9r " S S\5      rS/rg)z 
Processor class for Chameleon.
    )OptionalUnionN   )BatchFeature)
ImageInput)MultiModalDataProcessingKwargsProcessorMixin
TextKwargsUnpack)PreTokenizedInput	TextInputc                        \ rS rSr% \\S'   Srg)ChameleonTextKwargs#   return_for_text_completion N)__name__
__module____qualname____firstlineno__bool__annotations____static_attributes__r       j/var/www/html/shao/venv/lib/python3.13/site-packages/transformers/models/chameleon/processing_chameleon.pyr   r   #   s     $$r   r   F)totalc                   6    \ rS rSr% \\S'   SSSS.SS0S.rSrg	)
ChameleonProcessorKwargs'   text_kwargsF)paddingr   return_mm_token_type_idsreturn_tensorspt)r!   common_kwargsr   N)r   r   r   r   r   r   	_defaultsr   r   r   r   r   r   '   s+    $$ */(-
 d
	Ir   r   c                      ^  \ rS rSrSrSS/rSrSrSS\S\	4U 4S	 jjjr
    SS
\\   S\\\\\\   \\   4      S\\   S\4S jjrSS jrS rS r\S 5       rSrU =r$ )ChameleonProcessor5   a  
Constructs a Chameleon processor which wraps a Chameleon image processor and a Chameleon tokenizer into a single
processor.

[`ChameleonProcessor`] offers all the functionalities of [`ChameleonImageProcessor`] and [`LlamaTokenizerFast`].
See the [`~ChameleonProcessor.__call__`] and [`~ChameleonProcessor.decode`] for more information.

Args:
    image_processor ([`ChameleonImageProcessor`]):
        The image processor is a required input.
    tokenizer ([`LlamaTokenizerFast`]):
        The tokenizer is a required input.
    image_seq_length (`int`, *optional*, defaults to 1024):
        Sequence length of one image embedding.
    image_token (`str`, *optional*, defaults to `"<image>"`):
        The special token used to indicate image in the text.
image_processor	tokenizer)LlamaTokenizerLlamaTokenizerFastChameleonImageProcessorimage_seq_lengthimage_tokenc                 X  > X0l         [        US5      (       a  UR                  OUU l        UR                  U R                  5      U l        [        US5      (       a  UR
                  OSU l        [        US5      (       a  UR                  OSU l        UR                  U R                  5      U l        UR                  U R                  5      U l	        UR                  U R                  5      U l
        U R                  U R                  U R                  /U l        [        TU ]5  X5        g )Nr1   	boi_tokenz<racm3:break>	eoi_tokenz<eoss>)r0   hasattrr1   convert_tokens_to_idsimage_token_idr3   image_start_tokenr4   image_end_tokenimage_start_token_idimage_end_token_id	image_idssuper__init__)selfr+   r,   r0   r1   	__class__s        r   r>   ChameleonProcessor.__init__L   s     04;I}4U4U900[f'==d>N>NO#*9k#B#BI 	 7>i6U6Uy22[c'==d>N>NO$-$C$CDDZDZ$[!"+"A"A$BVBV"W--t/H/H$JaJab4r   imagestextkwargsreturnc                 8   [        U[        5      (       a  U/nO8[        U[        5      (       d#  [        US   [        5      (       d  [        S5      eUc  Uc  [	        S5      eU R
                  " [        4SU R                  R                  0UD6nUS   R                  SS5      n/ nU R                  U R                  U R                  -  -   U R                  -   n	U HO  n
U
R                  U R                  U	5      n
U(       d  XR                  R                  -  n
UR!                  U
5        MQ     0 nUb  U R"                  " U40 US	   D6nUS   R                  S
S5      nUS   R                  SS5      nU R                  " U40 US   DS
S0D6nU R%                  XS/S9  U(       ah  [&        R(                  " US   5      n[&        R*                  " US   5      nSU[&        R,                  " XR.                  5      '   UR1                  5       US'   [3        0 UEUEUS9$ )a  
Main method to prepare for the model one or several sequences(s) and image(s). This method forwards the `text`
and `kwargs` arguments to LlamaTokenizerFast's [`~LlamaTokenizerFast.__call__`] if `text` is not `None` to encode
the text. To prepare the image(s), this method forwards the `images` and `kwrags` arguments to
CLIPImageProcessor's [`~CLIPImageProcessor.__call__`] if `images` is not `None`. Please refer to the docstring
of the above two methods for more information.

Args:
    images (`PIL.Image.Image`, `np.ndarray`, `torch.Tensor`, `list[PIL.Image.Image]`, `list[np.ndarray]`, `list[torch.Tensor]`):
        The image or batch of images to be prepared. Each image can be a PIL image, NumPy array or PyTorch
        tensor. Both channels-first and channels-last formats are supported.
    text (`str`, `list[str]`, `list[list[str]]`):
        The sequence or batch of sequences to be encoded. Each sequence can be a string or a list of strings
        (pretokenized string). If the sequences are provided as list of strings (pretokenized), you must set
        `is_split_into_words=True` (to lift the ambiguity with a batch of sequences).
    return_tensors (`str` or [`~utils.TensorType`], *optional*):
        If set, will return tensors of a particular framework. Acceptable values are:

        - `'tf'`: Return TensorFlow `tf.constant` objects.
        - `'pt'`: Return PyTorch `torch.Tensor` objects.
        - `'np'`: Return NumPy `np.ndarray` objects.
        - `'jax'`: Return JAX `jnp.ndarray` objects.

Returns:
    [`BatchFeature`]: A [`BatchFeature`] with the following fields:

    - **input_ids** -- List of token ids to be fed to a model. Returned when `text` is not `None`.
    - **attention_mask** -- List of indices specifying which tokens should be attended to by the model (when
      `return_attention_mask=True` or if *"attention_mask"* is in `self.model_input_names` and if `text` is not
      `None`).
    - **pixel_values** -- Pixel values to be fed to a model. Returned when `images` is not `None`.
r   zAInvalid input text. Please provide a string, or a list of stringsNz&You must provide either text or imagestokenizer_init_kwargsr!   r   Fimages_kwargsr$   r#   image)
modalities	input_ids   mm_token_type_ids)datatensor_type)
isinstancestrlist	TypeError
ValueError_merge_kwargsr   r,   init_kwargspopr8   r1   r0   r9   replace	sep_tokenappendr+   _check_special_mm_tokensnparray
zeros_likeisinr<   tolistr   )r?   rB   rC   audiovideosrD   output_kwargsr   prompt_stringsone_img_tokenssampleimage_inputsr$   r#   text_inputs	array_idsrM   s                    r   __call__ChameleonProcessor.__call__[   s   R dC  6DD$''
47C0H0H_``<FNEFF**$
"&.."<"<
 

 &3=%A%E%EFbdi%j" //43C3CdF[F[3[\_c_s_ssF^^D$4$4nEF-..222!!&)	  //Y-:XYL&}599:JDQ#0#?#C#CD^`e#f nn^i}]7Sidhi%%nwi%X#[!9:I "k+.F GDEbggi@A/@/G/G/IK+,!@K!@<!@n]]r   c                     0 nUb>  U R                   S-   /[        U5      -  nS/[        U5      -  nUR                  XES.5        [        S0 UD6$ )a{  
Computes the number of placeholder tokens needed for multimodal inputs with the given sizes.

Args:
    image_sizes (`list[list[int]]`, *optional*):
        The input sizes formatted as (height, width) per each image.

Returns:
    `MultiModalData`: A `MultiModalData` object holding number of tokens per each of the provided
    input modalities, along with other useful data.
   rL   )num_image_tokensnum_image_patchesr   )r0   lenupdater   )r?   image_sizesrD   vision_datarn   ro   s         r   _get_num_multimodal_tokens-ChameleonProcessor._get_num_multimodal_tokens   s_     " $ 5 5 9:S=MM!"c+&6 64Dmn,,,r   c                 :    U R                   R                  " U0 UD6$ )z
This method forwards all its arguments to LlamaTokenizerFast's [`~PreTrainedTokenizer.batch_decode`]. Please
refer to the docstring of this method for more information.
)r,   batch_decoder?   argsrD   s      r   rw   ChameleonProcessor.batch_decode   s    
 ~~**D;F;;r   c                 :    U R                   R                  " U0 UD6$ )z
This method forwards all its arguments to LlamaTokenizerFast's [`~PreTrainedTokenizer.decode`]. Please refer to
the docstring of this method for more information.
)r,   decoderx   s      r   r|   ChameleonProcessor.decode   s    
 ~~$$d5f55r   c                     U R                   R                  nU R                  R                  n[        [        R                  X-   5      5      $ N)r,   model_input_namesr+   rR   dictfromkeys)r?   tokenizer_input_namesimage_processor_input_namess      r   r   $ChameleonProcessor.model_input_names   s>     !% @ @&*&:&:&L&L#DMM"7"UVWWr   )r9   r;   r<   r0   r8   r:   r1   r7   )i   z<image>)NNNNr   )r   r   r   r   __doc__
attributestokenizer_classimage_processor_classintrQ   r>   r   r   r   r   r   rR   r   r   r   rj   rt   rw   r|   propertyr   r   __classcell__)r@   s   @r   r)   r)   5   s    $ $[1J>O55S 5^a 5 5" (,hlO^$O^ uY(94	?DQbLccdeO^ 12O^ 
O^b-0<6 X Xr   r)   )r   typingr   r   numpyr\   feature_extraction_utilsr   image_utilsr   processing_utilsr   r	   r
   r   r   tokenization_utils_baser   r   r   r   r)   __all__r   r   r   <module>r      s`    #  4 %  D%*E %/u cX cXL  
 r   