Skip to main content

Apiche's group workspace

no_quant_vllm_2

What makes this group special?
Tags

no_quant_vllm_2/finetune

Notes
Author
State
Crashed
Start time
September 19th, 2025 9:28:19 PM
Runtime
3m 31s
Tracked hours
2m 53s
Run path
apiche/pipeline-rl/no_quant_vllm_2_finetune
OS
Linux-5.15.0-1067-nvidia-x86_64-with-glibc2.39
Python version
CPython 3.11.11
Git repository
git clone git@github.com:ServiceNow/pipelinerl.git
Git state
git checkout -b "no_quant_vllm_2/finetune" 88828596361fe13ad27b8fb8cc6cbb58dda41ce0
Command
pipelinerl/entrypoints/run_finetune.py --config-dir results/no_quant_vllm_2/conf --config-name exp_config output_dir=results/no_quant_vllm_2 hydra.run.dir=results/no_quant_vllm_2/finetune +me.weight_update_group_init_method=tcp://localhost:9000 +me.weight_update_group_world_size=3 +me.llm_urls=http://localhost:8080+http://localhost:8081
System Hardware
CPU count112
Logical CPU count 224
GPU count4
GPU typeNVIDIA H100 80GB HBM3
W&B CLI Version
0.19.11
Config

Config parameters are your model's inputs. Learn more

  • {} 186 keys
    • "False"
    • "no"
    • null
    • 1
    • 64
    • 0
    • 64
    • 64
    • "pipelinerl.domains.math.generate_math_rollout"
    • 1
    • 10,000,000
    • "Please reason step by step, and put your final answer within \boxed{}."
    • "{task}"
    • 50
    • 1
    • "nccl"
    • "pipelinerl.domains.math.load_datasets"
    • "False"
    • ""
    • true
    • null
    • false
    • "deepspeed_stage3_bf16"
    • "DeepSpeedPlugin(hf_ds_config=<accelerate.utils.deepspeed.HfDeepSpeedConfig object at 0x7ffcb3c0d4d0>, gradient_accumulation_steps=1, gradient_clipping='auto', zero_stage=3, is_train_batch_min=True, offload_optimizer_device='none', offload_param_device='none', offload_optimizer_nvme_path='none', offload_param_nvme_path='none', zero3_init_flag=True, zero3_save_16bit_model=True, transformer_moe_cls_names=None, enable_msamp=False, msamp_opt_level='O1')"
    • "cuda:0"
    • "DistributedType.DEEPSPEED"
    • "TorchDynamoPlugin(backend=<DynamoBackend.NO: 'NO'>, mode='default', fullgraph=False, dynamic=None, options=None, disable=False, use_regional_compilation=False)"
    • "pipelinerl.domains.math.MathEnvironment"
    • 78,000
    • [] 0 items
      • 1
      • "flash_attention_2"
      • false
      • "/mnt/llmd/base_models/Qwen2.5-0.5B"
      • true
      • null
      • "tapeagents.finetune.eval.dummy_eval_callback"
      • ""
      • true
      • 16
      • true
      • 0.3
      • "training_data"
      • -1
      • true
      • 0.000001
      • 46 ... 95
        96 ... 145
        146 ... 181
      • 7,777
      • 4
      • 0
      • 1
    Summary

    Summary metrics are your model's outputs. Learn more

    • {} 69 keys
      • -0.11091269552707672
      • 0
      • 0
      • 2.327451705932617
      • 0
      • 0.9985649915435408
      • 1,423.375
      • 0.001098044216632843
      • 0
      • -425.6705322265625
      • 0.984375
      • 0.3465912342071533
      • 50.93404769897461
      • 0
      • 0.0625
      • -0.734375
      • -0.00000011920928955078
      • -413.8691101074219
      • 0
      • 0.0625
      • -2.331266403198242
      • 0
      • 20,876
      • -2.3300094604492188
      • 4.988974094390869
      • 0.9998214244842528
      • 20,900.0625
      • 20,873.03515625
      • 1.0023549795150757
      • 1
      • -2.3300094604492188
      • 0
      • 0
      • 0
      • 0.0625
      • 36.634979248046875
      • 0.734375
      • 0.11091269552707672
      • -0.984375
      • 0.04870666563510895
      • 4
      • 0
      • 186.3104960790015
      • 0
      • 0.000001
      • 48
      • 46 ... 64
      • 3,893
      • 353.1244039261618
      • 7,786
      • 176.5622019630809
    Artifact Outputs

    This run produced these artifacts as outputs. Total: 1. Learn more