
    6hA                     P    S r SSKJrJrJrJrJrJrJrJ	r	J
r
JrJrJrJrJrJrJr  g)a\  
Utils shared by different modes of quantization (eager/graph)

This file is in the process of migration to `torch/ao/quantization`, and
is kept here for compatibility while the migration process is ongoing.
If you are adding a new entry/functionality, please, add it to the
`torch/ao/quantization/utils.py`, while adding an import statement
here.
    )activation_dtypeactivation_is_int8_quantized"activation_is_statically_quantizedcalculate_qmin_qmaxcheck_min_max_validget_combined_dictget_qconfig_dtypesget_qparam_dictget_quant_typeget_swapped_custom_module_classgetattr_from_fqnis_per_channelis_per_tensorweight_dtypeweight_is_quantizedweight_is_statically_quantizedN)__doc__torch.ao.quantization.utilsr   r   r   r   r   r   r	   r
   r   r   r   r   r   r   r   r        P/var/www/html/shao/venv/lib/python3.13/site-packages/torch/quantization/utils.py<module>r      s#       r   