Skip to main content

Kastan's group workspace

Aug-05__11:14

What makes this group special?
Tags

q_allFP32_gpt_8B_PP4_TP4_25d

Notes
Tags
Aug-05__11:14
BATCH_SIZE32
NUM_EPOCHS=3
NUM_MICRO_BATCHES=4
SLURM=513422
TP=4
WORLD_SIZE=32
Author
State
Crashed
Start time
August 5th, 2022 4:16:03 PM
Runtime
0s
Tracked hours
-
Run path
kastan/LLM-Distributed-Quantization/16x2ja9m
W&B CLI Version
0.13.0
Config

Config parameters are your model's inputs. Learn more

  • {} 22 keys
    • 32
    • 1
    • {} 1 key
      • "AMP_TYPE.NAIVE"
    • 4
    • 0.00015
    • "./quant_gpt2_2.5d_tp4_bs32_lr0.00015/"
    • {} 1 key
      • "titans.loss.lm_loss.gpt_lmloss.GPTLMLoss"
    • {} 7 keys
      • {} 4 keys
        • "torch.float32"
        • "torch.float32"
        • "torch.bfloat16"
        • "torch.float32"
      • 3
      • 4
      • {} 2 keys
        • 0.00015
        • 0.01
      • {} 2 keys
        • 4
        • {} 3 keys
          • 1
          • "2.5d"
          • 4
      • "titans.model.quant_gpt.quant_gpt.quant_gpt2_8B"
      • "titans.model.quant_gpt.quant_gpt.quant_gpt2_xl"
      • 1,024
      • "2.5d"
      • 4
      • 128
      • 50,304
      • 1
      • 0.01
    Summary

    Summary metrics are your model's outputs. Learn more

    No summary metrics saved for this run.

    Check the summary metrics documentation for more information.