Skip to main content
Reports
Created by
Created On
Last edited
AMP vs Custom Quantization
Accelerating multi-node Large Language Model training with per-layer selective quantization (e.g. FP32 -> FP16) of the transformer architecture.
0
2022-08-02