Pszemraj's workspace
Runs
34
Name
29 visualized
State
Tags
Notes
Killed
-
Crashed
-
Finished
-
Failed
-
Finished
-
Finished
finetune
fp32
-
Crashed
finetune
fp32
killed bc same as sdp
Finished
finetune
fp32
-
Finished
finetune
fp32
-
Finished
finetune
fp32
-
Finished
finetune
-
Killed
finetune
-
Failed
finetune
-
Failed
-
Killed
-
Failed
finetune
-
Failed
finetune
-
Failed
finetune
-
Finished
finetune
-
Finished
finetune
-
Runtime
Hostname
GPU Type
GPU Count
_name_or_path
adafactor
adam_beta1
adam_beta2
adam_epsilon
add_cross_attention
architectures
attention_bias
attention_dropout
auto_find_batch_size
bench_dataset
bench_source_max_len
bench_split
bf16
bf16_full_eval
bos_token_id
chunk_size_feed_forward
data_seed
dataloader_drop_last
dataloader_num_workers
dataloader_persistent_workers
dataloader_pin_memory
ddp_find_unused_parameters
ddp_timeout
debug
disable_tqdm
diversity_penalty
do_bench_eval
do_eval
do_predict
do_sample
do_train
early_stopping
encoder_no_repeat_ngram_size
eos_token_id
eval_accumulation_steps
eval_delay
eval_sample_packing
eval_steps
evaluation_strategy
fp16
fp16_backend
fp16_full_eval
fp16_opt_level
fsdp
fsdp_config.min_num_params
fsdp_config.xla
7m 45s
-
NVIDIA RTX 6000 Ada Generation
1
TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
29889
false
32
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
30m 30s
-
Tesla T4
1
BEE-spoke-data/smol_llama-220M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
8
0
false
0.083333
steps
false
auto
false
O1
[]
0
false
5m 13s
-
Tesla T4
1
BEE-spoke-data/smol_llama-220M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
8
0
false
0.083333
steps
false
auto
false
O1
[]
0
false
20s
-
Tesla T4
1
./smol_llama-220M-GQA-local
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
4
0
false
0.083333
steps
true
auto
false
O1
[]
0
false
3m 57s
-
Tesla T4
1
BEE-spoke-data/smol_llama-220M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
2
0
false
0.083333
steps
false
auto
false
O1
[]
0
false
2m 51s
-
Tesla T4
1
BEE-spoke-data/smol_llama-101M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
4
0
false
0.125
steps
false
auto
false
O1
[]
0
false
18m 30s
-
Tesla T4
1
BEE-spoke-data/smol_llama-101M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
4
0
false
0.125
steps
false
auto
false
O1
[]
0
false
35m 59s
-
Tesla T4
1
BEE-spoke-data/smol_llama-101M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
4
0
false
0.125
steps
false
auto
false
O1
[]
0
false
11m 32s
-
Tesla T4
1
BEE-spoke-data/smol_llama-101M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
4
0
true
0.125
steps
false
auto
false
O1
[]
0
false
47m 59s
-
Tesla T4
1
BEE-spoke-data/smol_llama-101M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
false
false
1
0
-
false
0
false
true
-
1800
[]
false
0
false
true
false
false
false
false
0
2
4
0
true
0.083333
steps
false
auto
false
O1
[]
0
false
2h 15m 19s
-
NVIDIA GeForce RTX 4090
1
pszemraj/tinyllama-1.1b-3T
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
25978
false
7
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
5m 45s
-
NVIDIA GeForce RTX 4090
1
pszemraj/tinyllama-1.1b-3T
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
15476
false
7
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
30m 11s
-
NVIDIA GeForce RTX 4090
1
BEE-spoke-data/smol_llama-220M-GQA
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
7312
false
7
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
55m 10s
-
NVIDIA GeForce RTX 4090
1
BEE-spoke-data/llama-GQA-400m
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
28069
false
7
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
9m 51s
-
NVIDIA GeForce RTX 4090
1
BEE-spoke-data/llama-GQA-400m
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
3490
false
7
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
1h 1m 56s
-
NVIDIA GeForce RTX 4090
1
BEE-spoke-data/llama-GQA-400m
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
6992
false
7
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
10m 55s
-
NVIDIA GeForce RTX 4090
1
pszemraj/tinyllama-1.1b-3T
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
9661
false
7
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
7m 24s
-
NVIDIA GeForce RTX 4090
1
pszemraj/tinyllama-1.1b-3T
false
0.9
0.95
1.0000e-8
false
["LlamaForCausalLM"]
false
0
false
-
-
-
true
false
1
0
6306
false
7
false
true
-
1800
[]
false
0
-
true
false
false
true
false
0
2
-
0
-
50
steps
false
auto
false
O1
[]
0
false
17m 8s
-
NVIDIA GeForce RTX 3090
2
BEE-spoke-data/smol_llama-220M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
true
false
1
0
-
false
0
false
true
false
1800
[]
false
0
false
true
false
false
false
false
0
2
2
0
true
0.083333
steps
false
auto
false
O1
[]
0
false
1h 3m 54s
-
NVIDIA GeForce RTX 3090
2
BEE-spoke-data/smol_llama-220M-GQA
false
0.9
0.95
1.0000e-9
false
["LlamaForCausalLM"]
false
0
false
pharaouk/dharma-1/dharma_1_mini.json
2048
eval
true
false
1
0
-
false
0
false
true
false
1800
[]
false
0
false
true
false
false
false
false
0
2
2
0
true
0.083333
steps
false
auto
false
O1
[]
0
false
1-20
of 34