A0970601776's workspace
Runs
9
Name
4 visualized
State
Notes
User
Tags
Created
Runtime
Sweep
_attn_implementation_autoset
_name_or_path
accelerator_config.even_batches
accelerator_config.non_blocking
accelerator_config.split_batches
accelerator_config.use_seedable_sampler
adafactor
adam_beta1
adam_beta2
adam_epsilon
add_cross_attention
architectures
attention_bias
attention_dropout
auto_find_batch_size
batch_eval_metrics
bf16
bf16_full_eval
bos_token_id
chunk_size_feed_forward
dataloader_drop_last
dataloader_num_workers
dataloader_persistent_workers
dataloader_pin_memory
ddp_find_unused_parameters
ddp_timeout
debug
deepspeed
disable_tqdm
diversity_penalty
do_eval
do_predict
do_sample
do_train
early_stopping
encoder_no_repeat_ngram_size
eos_token_id
eval_delay
eval_do_concat_batches
eval_on_start
eval_strategy
eval_use_gather_object
fp16
fp16_backend
Finished
-
a0970601776
H100*8
adamw_torch_fused
batch_14
cutoff_len_1024
epochs_30
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
9h 29m 11s
-
true
meta-llama/Llama-3.1-8B
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm//LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
128001
0
true
false
no
false
false
auto
Crashed
-
a0970601776
H100*8
adamw_torch_fused
batch_14
cutoff_len_1024
epochs_30
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
2h 59m 16s
-
true
meta-llama/Llama-3.1-8B
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm//LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
128001
0
true
false
no
false
false
auto
Finished
-
a0970601776
H100*8
adamw_torch_fused
batch_14
cutoff_len_1024
epochs_3
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
24m 47s
-
true
meta-llama/Llama-3.1-8B
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm//LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
128001
0
true
false
no
false
false
auto
Finished
-
a0970601776
H100*8
adamw_torch_fused
batch_5
cutoff_len_1024
epochs_1
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
21m 59s
-
true
/mnt/llm/LLaMA-Factory/saves/Llama-3.1-Freego-8B-Instruct/sft/sft-2024-12-12-1/
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm//LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
128001
0
true
false
no
false
false
auto
Finished
-
a0970601776
H100*8
adamw_torch_fused
batch_5
cutoff_len_1024
epochs_1
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
21m 55s
-
true
/mnt/llm/LLaMA-Factory/saves/Llama-3.1-Freego-8B-Instruct/sft/sft-2024-12-12-1/
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm//LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
128001
0
true
false
no
false
false
auto
Crashed
-
a0970601776
H100*8
adamw_torch_fused
batch_5
cutoff_len_1024
epochs_10
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
46m 46s
-
true
/mnt/llm/LLaMA-Factory/saves/Llama-3.1-Freego-8B-Instruct/sft/sft-2024-12-12-1/
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm//LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
128001
0
true
false
no
false
false
auto
Finished
-
a0970601776
H100*8
adamw_torch_fused
batch_14
cutoff_len_1024
epochs_10
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
1h 10m 40s
-
true
meta-llama/Llama-3.1-8B
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm//LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
128001
0
true
false
no
false
false
auto
Crashed
-
a0970601776
H100*8
adamw_torch_fused
batch_5
cutoff_len_1024
epochs_10
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
27m 16s
-
true
meta-llama/Llama-3.1-8B
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm/LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
128001
0
true
false
no
false
false
auto
Crashed
-
a0970601776
H100*8
adamw_torch_fused
batch_5
cutoff_len_1024
epochs_10
grad_acc_8
lr_5e-6
warmup_ratio_0.01
z3
1h 46m 51s
-
true
meta-llama/Llama-3.1-8B-Instruct
true
false
false
true
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
false
true
false
128000
0
false
72
false
true
false
10800000
[]
/mnt/llm/LLaMA-Factory/examples/deepspeed/ds_z3_config.json
false
0
false
false
false
true
false
0
[128001,128008,128009]
0
true
false
no
false
false
auto
1-9
of 9