Skip to main content

Preetham-gali's group workspace

cawcJ6TCRPjouBa97Rtb3K_3jx6l7yy

What makes this group special?
Tags

sweep-agent-1-0-0

Notes
State
Finished
Start time
August 30th, 2021 12:44:06 PM
Runtime
4h 53m 57s
Tracked hours
4h 53m 51s
Run path
eleutherai/distilling/139l7vgg
OS
Linux-5.8.0-59-generic-x86_64-with-glibc2.29
Python version
3.8.5
Git repository
git clone https://github.com/EleutherAI/gpt-neox.git
Git state
git checkout -b "sweep-agent-1-0-0" e9b491a80df1dbb887dcd43adc7e8ec22c563a52
Command
pretrain_gpt2.py --local_rank=0 --num_gpus 8 --deepspeed_config "{\"train_batch_size\": 192, \"train_micro_batch_size_per_gpu\": 1, \"gradient_accumulation_steps\": 24, \"optimizer\": {\"type\": \"Adam\", \"params\": {\"lr\": 0.0003, \"betas\": [0.9, 0.999], \"eps\": 1e-08}}, \"fp16\": {\"enabled\": true, \"loss_scale\": 0, \"loss_scale_window\": 1000, \"hysteresis\": 2, \"min_loss_scale\": 1}, \"gradient_clipping\": 1.0, \"zero_optimization\": {\"stage\": 1, \"allgather_partitions\": true, \"allgather_bucket_size\": 500000000, \"overlap_comm\": true, \"reduce_scatter\": true, \"reduce_bucket_size\": 500000000, \"contiguous_gradients\": true, \"cpu_offload\": false}, \"wall_clock_breakdown\": true}" --megatron_config "{\"num_gpus\": 8, \"train_batch_size\": 192, \"train_micro_batch_size_per_gpu\": 1, \"gradient_accumulation_steps\": 24, \"optimizer\": {\"type\": \"Adam\", \"params\": {\"lr\": 0.0003, \"betas\": [0.9, 0.999], \"eps\": 1e-08}}, \"fp16\": {\"enabled\": true, \"loss_scale\": 0, \"loss_scale_window\": 1000, \"hysteresis\": 2, \"min_loss_scale\": 1}, \"gradient_clipping\": 1.0, \"zero_optimization\": {\"stage\": 1, \"allgather_partitions\": true, \"allgather_bucket_size\": 500000000, \"overlap_comm\": true, \"reduce_scatter\": true, \"reduce_bucket_size\": 500000000, \"contiguous_gradients\": true, \"cpu_offload\": false}, \"wall_clock_breakdown\": true, \"precision\": \"fp16\", \"lr_decay_style\": \"cosine\", \"lr_decay_iters\": 250000, \"optimizer_type\": \"Adam\", \"zero_stage\": 1, \"zero_reduce_scatter\": true, \"zero_contiguous_gradients\": true, \"zero_reduce_bucket_size\": 500000000, \"zero_allgather_bucket_size\": 500000000, \"lr\": 0.0003, \"data_path\": \"/mnt/ssd-1/data/pile/pile_text_document\", \"data_impl\": \"mmap\", \"save\": \"checkpoints/distilling/large-to-med1\", \"load\": \"checkpoints/distilling/large-to-med1\", \"save_interval\": 10000, \"finetune\": true, \"batch_size\": 1, \"train_iters\": 250000, \"eval_iters\": 10, \"keep_last_n_checkpoints\": 4, \"split\": \"949,50,1\", \"vocab_file\": \"/mnt/ssd-1/data/gpt2-vocab.json\", \"merge_file\": \"/mnt/ssd-1/data/gpt2-merges.txt\", \"attention_dropout\": 0.0, \"hidden_dropout\": 0.0, \"weight_decay\": 0.1, \"checkpoint_activations\": true, \"synchronize_each_layer\": true, \"partition_activations\": true, \"gas\": 24, \"clip_grad\": 1.0, \"dynamic_loss_scale\": true, \"do_distillation\": true, \"teacher_model_args\": {\"precision\": null, \"num_layers\": 24, \"hidden_size\": 1536, \"num_attention_heads\": 16, \"seq_length\": 2048, \"max_position_embeddings\": 2048, \"norm\": \"layernorm\", \"layernorm_epsilon\": 1e-05, \"rms_norm_epsilon\": 1e-08, \"scalenorm_epsilon\": 1e-08, \"pos_emb\": \"rotary\", \"rpe_num_buckets\": 32, \"rpe_max_distance\": 128, \"no_weight_tying\": true, \"attention_config\": [\"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\"], \"sparsity_config\": {}, \"num_unique_layers\": null, \"param_sharing_style\": \"grouped\", \"make_vocab_size_divisible_by\": 128, \"apply_residual_connection_post_layernorm\": false, \"activation\": \"gelu\", \"scaled_upper_triang_masked_softmax_fusion\": false, \"scaled_masked_softmax_fusion\": false, \"bias_gelu_fusion\": false, \"bias_dropout_fusion\": false, \"fp16_lm_cross_entropy\": false, \"init_method_std\": 0.02, \"apply_query_key_layer_scaling\": false, \"use_cpu_initialization\": false, \"attention_softmax_in_fp32\": false, \"rotary_pct\": 1.0, \"rotary_emb_base\": 10000, \"init_method\": \"normal\", \"output_layer_init_method\": \"scaled_normal\", \"gmlp_attn_dim\": 64}, \"student_model_args\": {\"precision\": null, \"num_layers\": 24, \"hidden_size\": 1024, \"num_attention_heads\": 16, \"seq_length\": 2048, \"max_position_embeddings\": 2048, \"norm\": \"layernorm\", \"layernorm_epsilon\": 1e-05, \"rms_norm_epsilon\": 1e-08, \"scalenorm_epsilon\": 1e-08, \"pos_emb\": \"rotary\", \"rpe_num_buckets\": 32, \"rpe_max_distance\": 128, \"no_weight_tying\": true, \"attention_config\": [\"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\", \"global\"], \"sparsity_config\": {}, \"num_unique_layers\": null, \"param_sharing_style\": \"grouped\", \"make_vocab_size_divisible_by\": 128, \"apply_residual_connection_post_layernorm\": false, \"activation\": \"gelu\", \"scaled_upper_triang_masked_softmax_fusion\": false, \"scaled_masked_softmax_fusion\": false, \"bias_gelu_fusion\": false, \"bias_dropout_fusion\": false, \"fp16_lm_cross_entropy\": false, \"init_method_std\": 0.02, \"apply_query_key_layer_scaling\": false, \"use_cpu_initialization\": false, \"attention_softmax_in_fp32\": false, \"rotary_pct\": 1.0, \"rotary_emb_base\": 10000, \"init_method\": \"normal\", \"output_layer_init_method\": \"scaled_normal\", \"gmlp_attn_dim\": 64}, \"load_teacher\": \"/mnt/ssd-1/neox_checkpoints/dense_large_checkpoints/global_step250000\", \"alpha_lm\": 1.0, \"alpha_kld\": 0.5, \"pipe_parallel_size\": 1, \"is_pipe_parallel\": true, \"use_wandb\": true, \"wandb_group\": \"cawcJ6TCRPjouBa97Rtb3K_3jx6l7yy\", \"wandb_team\": \"eleutherai\", \"wandb_project\": \"distilling\", \"log_dir\": \"logs\", \"log_interval\": 100, \"user_script\": \"pretrain_gpt2.py\"}"
System Hardware
CPU count96
GPU count8
GPU typeNVIDIA A100-PCIE-40GB
W&B CLI Version
0.12.1
Config

Config parameters are your model's inputs. Learn more

  • {} 186 keys
    • "gelu"
    • false
    • 1,000
    • 0.5
    • 1
    • 0
    • null
    • false
    • false
    • null
    • 0
    • false
    • 1
    • false
    • false
    • true
    • false
    • 1
    • false
    • 1
    • false
    • "mmap"
    • "/mnt/ssd-1/data/pile/pile_text_document"
    • false
    • null
    • true
    • true
    • false
    • false
    • "nccl"
    • true
    • null
    • null
    • null
    • false
    • true
    • false
    • 1,000
    • 10
    • ""
    • null
    • null
    • null
    • true
    • null
    • {} 5 keys
      • 46 ... 95
        96 ... 145
        146 ... 181
      • {} 8 keys
        • 500,000,000
        • true
        • 1
      Summary

      Summary metrics are your model's outputs. Learn more

      • {} 38 keys
        • 99,002,917,324,128.94
        • 5.192514305114746
        • 36.97630641303686
        • 0.3116130828857422
        • 0.12874603271484375
        • 277.0719528198242
        • 0.8978843688964844
        • 23,145.246744155884
        • 2.748727798461914
        • 7.930755615234375
        • 23,122.793197631836
        • 23,128.954887390137
        • 23,150.82550048828
        • 348.21248054504395
        • 2,548.844337463379
        • 20,176.156520843502
        • 45.39121730999124
        • 4.998656894437888
        • 39.56839670975842
        • 0.5472861763086839
        • 2,360.4211807250977
        • 0.4017353057861328
        • 279.064416885376
        • 0
        • NaN
        • 0.0002999917226836827
        • 9.9374361038208
        • NaN
        • 1
        • 0
        • 485,165,195.4097903
        • 9.32800579071045
        • 11,248.676887437923
        • 485,165,195.4097903
        • 0
        • 1
      Artifact Outputs

      This run produced these artifacts as outputs. Total: 1. Learn more

      Loading...