# You can save the transformer weights as bf16 or 8-bit with the --do_8_bit flag # You can also save with scaled 8-bit using the --do_8bit_scaled flag # Call like this for 8-bit transformer weights ...
from mistral_common.tokens.tokenizers.base import TokenizerVersion # pyright: ignore[reportMissingImports] from mistral_common.tokens.tokenizers.multimodal import DATASET_MEAN as ...