BitsandbytesPrecision¶
- class lightning.pytorch.plugins.precision.BitsandbytesPrecision(mode, dtype=None, ignore_modules=None)[source]¶
- Bases: - Precision,- BitsandbytesPrecision- Plugin for quantizing weights with bitsandbytes. - Warning - This is an experimental feature. - Note - The optimizer is not automatically replaced with - bitsandbytes.optim.Adam8bitor equivalent 8-bit optimizers.- Parameters:
- mode¶ ( - Literal[- 'nf4',- 'nf4-dq',- 'fp4',- 'fp4-dq',- 'int8',- 'int8-training']) – The quantization mode to use.
- ignore_modules¶ ( - Optional[- set[- str]]) – The submodules whose Linear layers should not be replaced, for example.- {"lm_head"}. This might be desirable for numerical stability. The string will be checked in as a prefix, so a value like “transformer.blocks” will ignore all linear layers in all of the transformer blocks.