Skip to main content
kastan
Projects
LLM-Distributed-Quantization
Reports
Log in
Sign up
Project
Workspace
Runs
Automat.
Sweeps
Reports
Artifacts
Anyone
Anyone
kastan
Reports
Created by
Created On
Last edited
AMP vs Custom Quantization
Accelerating multi-node Large Language Model training with per-layer selective quantization (e.g. FP32 -> FP16) of the transformer architecture.
0
kastan
2022-08-02
3 years ago
Clone report