Apiche's group workspace
debug_actor_test_logging_july_23rd_1_guessing
What makes this group special?
Tags
debug_actor_test_logging_july_23rd_1_guessing/finetune
Notes
Author
State
Crashed
Start time
July 23rd, 2025 2:53:27 PM
Runtime
5m 31s
Tracked hours
5m 9s
Run path
apiche/pipeline-rl/debug_actor_test_logging_july_23rd_1_guessing_finetune
OS
Linux-5.15.0-1067-nvidia-x86_64-with-glibc2.39
Python version
CPython 3.11.11
Git repository
git clone git@github.com:ServiceNow/pipelinerl.git
Git state
git checkout -b "debug_actor_test_logging_july_23rd_1_guessing/finetune" 3a6aca9a40fe429fafb5d0da5edb0450a11cb497
Command
pipelinerl/entrypoints/run_finetune.py --config-dir results/debug_actor_test_logging_july_23rd_1_guessing/conf --config-name exp_config output_dir=results/debug_actor_test_logging_july_23rd_1_guessing hydra.run.dir=results/debug_actor_test_logging_july_23rd_1_guessing/finetune +me.weight_update_group_init_method=tcp://localhost:9000 +me.weight_update_group_world_size=2 +me.llm_urls=http://localhost:8080
System Hardware
| CPU count | 112 |
| Logical CPU count | 224 |
| GPU count | 4 |
| GPU type | NVIDIA H100 80GB HBM3 |
W&B CLI Version
0.19.11
Config
Config parameters are your model's inputs. Learn more
- {} 181 keys▶
- "False"
- "no"
- null
- 1
- 64
- 0
- 64
- 64
- "pipelinerl.domains.guessing.generate_guessing_rollout"
- 1
- 10,000,000
- 50
- 1
- "nccl"
- "pipelinerl.domains.guessing.load_problems"
- "False"
- ""
- true
- null
- false
- "deepspeed_stage3_bf16"
- "DeepSpeedPlugin(hf_ds_config=<accelerate.utils.deepspeed.HfDeepSpeedConfig object at 0x7ffc75c1a3d0>, gradient_accumulation_steps=1, gradient_clipping='auto', zero_stage=3, is_train_batch_min=True, offload_optimizer_device='none', offload_param_device='none', offload_optimizer_nvme_path='none', offload_param_nvme_path='none', zero3_init_flag=True, zero3_save_16bit_model=True, transformer_moe_cls_names=None, enable_msamp=False, msamp_opt_level='O1')"
- "cuda:0"
- "DistributedType.DEEPSPEED"
- "TorchDynamoPlugin(backend=<DynamoBackend.NO: 'NO'>, mode='default', fullgraph=False, dynamic=False, options=None, disable=False)"
- null
- 20,000
- [] 0 items
- "flash_attention_2"
- false
- "Qwen/Qwen2.5-7B-Instruct"
- true
- null
- "tapeagents.finetune.eval.dummy_eval_callback"
- ""
- true
- 1,026
- true
- 0.3
- "training_data"
- -1
- true
- 0.000001
- true
- 1
- 16
- 7,777
- 6
- 0
- 1
46 ... 95▶▶96 ... 145▶▶146 ... 176▶▶
Summary
Summary metrics are your model's outputs. Learn more
- {} 69 keys▶
- -1.529497146606445
- 0
- 0
- 0.07191143184900284
- 0
- 0.996103138464522
- 120.97270965576172
- 0.004619069863110781
- 0
- 0.3180975914001465
- 1.807812452316284
- 0.9836704730987548
- 0.02510165050625801
- 1.2999999523162842
- 0.000974658876657486
- -2.4585938453674316
- -0.00000005960464477539
- -0.019365161657333377
- -1.899999976158142
- 0.000974658876657486
- -0.08967616409063339
- 0
- 9,486
- -0.08333136886358261
- 0.12415313720703124
- 0.9969342947006226
- 9,523.6376953125
- 9,486.2626953125
- 1.010963797569275
- 1
- -0.08333136886358261
- -1.3786550760269165
- 0
- 0
- 0.000974658876657486
- 0.011088640429079533
- 0.71875
- 0.15084204077720642
- -0.67578125
- 2.4945549964904785
- 15
- 0
- 0.21287604865756085
- 1,026
- 0.00000028
- 14,364
- 3,829.909090909091
- 8,303.85229503267
- 126,387
- 2,767.95076501089
46 ... 64▶▶
Artifact Outputs
This run produced these artifacts as outputs. Total: 1. Learn more
Type
Name
Consumer count
Loading...