Skip to main content

Chilli's group workspace

G4NuteJNEsu5fhd2Htcs6a

What makes this group special?
Tags

neox-kip-0-7

Notes
State
Crashed
Start time
March 26th, 2021 3:10:41 PM
Runtime
53m 4s
Tracked hours
-
Run path
eleutherai/neox/w3sfgi8r
OS
Linux-5.8.0-45-generic-x86_64-with-glibc2.29
Python version
3.8.5
Git repository
git clone https://github.com/EleutherAI/gpt-neox.git
Git state
git checkout -b "neox-kip-0-7" cd0f0b0ce92acf44f5a2cb87c3b35e098c62c896
Command
pretrain_gpt2.py --local_rank=7 --num-layers 24 --hidden-size 1024 --num-attention-heads 16 --max-position-embeddings 2048 --attention-dropout 0 --hidden-dropout 0 --weight-decay 0 --batch-size 4 --checkpoint-activations --checkpoint-num-layers 1 --train-iters 320000 --log-interval 100 --tensorboard-dir /mnt/ssd-cluster/tensorboard --pos-emb none --norm rmsnorm --lr-decay-style cosine --lr-decay-iters 320000 --warmup 0.01 --save /mnt/ssd-cluster/checkpoints --save-interval 10000 --keep-last-n-checkpoints 4 --model-parallel-size 1 --pipe-parallel-size 0 --distributed-backend nccl --eval-iters 10 --eval-interval 1000 --data-path /mnt/ssd-cluster/data/enron/enron_text_document --split 949,50,1 --vocab-file /mnt/ssd-cluster/data/gpt2-vocab.json --merge-file /mnt/ssd-cluster/data/gpt2-merges.txt --seq-length 2048 --data-impl mmap --log-dir /mnt/ssd-cluster/logs --partition-activations --synchronize-each-layer --wandb_group G4NuteJNEsu5fhd2Htcs6a --wandb_team eleutherai --git_hash cd0f0b0 --deepspeed --fp16 --gas 1 --zero-stage 1 --zero-reduce-scatter --zero-contiguous-gradients --zero-reduce-bucket-size 500000000 --zero-allgather-bucket-size 500000000 --clip-grad 1.0 --lr 0.0003 --adam-beta1 0.9 --adam-beta2 0.95 --adam-eps 1e-08 --momentum 0.0 --deepspeed_config {"train_batch_size":32.0,"train_micro_batch_size_per_gpu":4,"gradient_accumulation_steps":1,"optimizer":{"type":"Adam","params":{"lr":0.0003,"max_grad_norm":1.0,"betas":[0.9,0.95]}},"fp16":{"fp16":true,"enabled":true,"loss_scale":0,"loss_scale_window":1000,"hysteresis":2,"min_loss_scale":1},"gradient_clipping":1.0,"zero_optimization":{"stage":1,"allgather_partitions":true,"allgather_bucket_size":500000000,"overlap_comm":true,"reduce_scatter":true,"reduce_bucket_size":500000000,"contiguous_gradients":true,"cpu_offload":false},"steps_per_print":10,"wall_clock_breakdown":true,"deepspeed":true}
System Hardware
CPU count40
GPU count8
GPU typeGeForce RTX 2080 Ti
W&B CLI Version
0.10.21
Config

Config parameters are your model's inputs. Learn more

  • {} 121 keys
    • 0.9
    • 0.95
    • 0.00000001
    • false
    • 1,000
    • false
    • false
    • 0
    • false
    • 4
    • false
    • false
    • true
    • false
    • 1
    • 1
    • false
    • false
    • false
    • "mmap"
    • "/mnt/ssd-cluster/data/enron/enron_text_document"
    • false
    • null
    • true
    • false
    • "{"train_batch_size":32.0,"train_micro_batch_size_per_gpu":4,"gradient_accumulation_steps":1,"optimizer":{"type":"Adam","params":{"lr":0.0003,"max_grad_norm":1.0,"betas":[0.9,0.95]}},"fp16":{"fp16":true,"enabled":true,"loss_scale":0,"loss_scale_window":1000,"hysteresis":2,"min_loss_scale":1},"gradient_clipping":1.0,"zero_optimization":{"stage":1,"allgather_partitions":true,"allgather_bucket_size":500000000,"overlap_comm":true,"reduce_scatter":true,"reduce_bucket_size":500000000,"contiguous_gradients":true,"cpu_offload":false},"steps_per_print":10,"wall_clock_breakdown":true,"deepspeed":true}"
    • false
    • false
    • "nccl"
    • true
    • false
    • 1,000
    • 10
    • null
    • false
    • true
    • false
    • false
    • 1
    • false
    • "cd0f0b0"
    • 0
    • 1,024
    • 2
    • 0.02
    • 4
    • 46 ... 95
      96 ... 116
    • true
    • 500,000,000
    • true
    • 1
Summary

Summary metrics are your model's outputs. Learn more

No summary metrics saved for this run.

Check the summary metrics documentation for more information.