Chilli's group workspace
SWDKQ6x8Rs5E2xzRZ9NxiV
What makes this group special?
Tags
neox-kip-0-3
Notes
Author
State
Crashed
Start time
March 26th, 2021 4:05:04 PM
Runtime
49m 47s
Tracked hours
-
Run path
eleutherai/neox/ta7fi6s6
OS
Linux-5.8.0-45-generic-x86_64-with-glibc2.29
Python version
3.8.5
Git repository
git clone https://github.com/EleutherAI/gpt-neox.git
Git state
git checkout -b "neox-kip-0-3" cd0f0b0ce92acf44f5a2cb87c3b35e098c62c896
Command
pretrain_gpt2.py --local_rank=3 --num-layers 24 --hidden-size 1024 --num-attention-heads 16 --max-position-embeddings 2048 --attention-dropout 0 --hidden-dropout 0 --weight-decay 0 --batch-size 4 --checkpoint-activations --checkpoint-num-layers 1 --train-iters 320000 --log-interval 100 --tensorboard-dir /mnt/ssd-cluster/tensorboard --pos-emb none --norm rmsnorm --lr-decay-style cosine --lr-decay-iters 320000 --warmup 0.01 --save /mnt/ssd-cluster/checkpoints --save-interval 10000 --keep-last-n-checkpoints 4 --model-parallel-size 1 --pipe-parallel-size 0 --distributed-backend nccl --eval-iters 10 --eval-interval 1000 --data-path /mnt/ssd-cluster/data/enron/enron_text_document --split 949,50,1 --vocab-file /mnt/ssd-cluster/data/gpt2-vocab.json --merge-file /mnt/ssd-cluster/data/gpt2-merges.txt --seq-length 2048 --data-impl mmap --log-dir /mnt/ssd-cluster/logs --partition-activations --synchronize-each-layer --wandb_group SWDKQ6x8Rs5E2xzRZ9NxiV --wandb_team eleutherai --git_hash cd0f0b0 --deepspeed --fp16 --gas 1 --zero-stage 0 --zero-reduce-scatter --zero-contiguous-gradients --zero-reduce-bucket-size 500000000 --zero-allgather-bucket-size 500000000 --clip-grad 1.0 --lr 0.0003 --adam-beta1 0.9 --adam-beta2 0.95 --adam-eps 1e-08 --momentum 0.0 --deepspeed_config {"train_batch_size":32.0,"train_micro_batch_size_per_gpu":4,"gradient_accumulation_steps":1,"optimizer":{"type":"Adam","params":{"lr":0.0003,"max_grad_norm":1.0,"betas":[0.9,0.95]}},"fp16":{"fp16":true,"enabled":true,"loss_scale":0,"loss_scale_window":1000,"hysteresis":2,"min_loss_scale":1},"gradient_clipping":1.0,"zero_optimization":{"stage":0,"allgather_partitions":true,"allgather_bucket_size":500000000,"overlap_comm":true,"reduce_scatter":true,"reduce_bucket_size":500000000,"contiguous_gradients":true,"cpu_offload":false},"steps_per_print":10,"wall_clock_breakdown":true,"deepspeed":true}
System Hardware
CPU count | 40 |
GPU count | 8 |
GPU type | GeForce RTX 2080 Ti |
W&B CLI Version
0.10.21
Config
Config parameters are your model's inputs. Learn more
- {} 121 keys▶
- 0.9
- 0.95
- 0.00000001
- false
- 1,000
- false
- false
- 0
- false
- 4
- false
- false
- true
- false
- 1
- 1
- false
- false
- false
- "mmap"
- "/mnt/ssd-cluster/data/enron/enron_text_document"
- false
- null
- true
- false
- "{"train_batch_size":32.0,"train_micro_batch_size_per_gpu":4,"gradient_accumulation_steps":1,"optimizer":{"type":"Adam","params":{"lr":0.0003,"max_grad_norm":1.0,"betas":[0.9,0.95]}},"fp16":{"fp16":true,"enabled":true,"loss_scale":0,"loss_scale_window":1000,"hysteresis":2,"min_loss_scale":1},"gradient_clipping":1.0,"zero_optimization":{"stage":0,"allgather_partitions":true,"allgather_bucket_size":500000000,"overlap_comm":true,"reduce_scatter":true,"reduce_bucket_size":500000000,"contiguous_gradients":true,"cpu_offload":false},"steps_per_print":10,"wall_clock_breakdown":true,"deepspeed":true}"
- false
- false
- "nccl"
- true
- false
- 1,000
- 10
- null
- false
- true
- false
- false
- 1
- false
- "cd0f0b0"
- 0
- 1,024
- 2
- 0.02
- 4
- true
- 500,000,000
- true
- 0
46 ... 95▶▶96 ... 116▶▶
Summary
Summary metrics are your model's outputs. Learn more
No summary metrics saved for this run.
Check the summary metrics documentation for more information.