Autometa's workspace
Runs
38
State
Notes
User
Tags
Created
Runtime
Sweep
_name_or_path
accelerator_config.even_batches
accelerator_config.split_batches
accelerator_config.use_seedable_sampler
adafactor
adam_beta1
adam_beta2
adam_epsilon
add_cross_attention
architectures
attention_bias
attention_dropout
auto_find_batch_size
beta
bf16
bf16_full_eval
bos_token_id
chunk_size_feed_forward
cli_configs.batch_size
cli_configs.batch_sizes
cli_configs.bootstrap_iters
cli_configs.device
cli_configs.limit
cli_configs.model
cli_configs.model_args
dataloader_drop_last
dataloader_num_workers
dataloader_persistent_workers
dataloader_pin_memory
ddp_timeout
debug
disable_tqdm
diversity_penalty
do_eval
do_predict
do_sample
do_train
early_stopping
encoder_no_repeat_ngram_size
eos_token_id
eval_delay
eval_steps
evaluation_strategy
fp16
Crashed
capecape
5d 7h 22m 49s
-
["/workspace/artifacts/Mistral-7B-v0.1:v1","/workspace/artifacts/Mistral-7B-v0.1:v2","/workspace/artifacts/Mistral-7B-v0.1:v3","/workspace/artifacts/Mistral-7B-v0.1:v6","/workspace/artifacts/Mistral-7B-v0.1:v7","mistralai/Mistral-7B-v0.1"]
true
false
true
false
0.9
0.999
1.0000e-8
false
MistralForCausalLM
-
0
false
0.05
true
false
1
0
-
-
-
-
-
-
-
false
0
false
true
1800
-
false
0
true
false
false
false
false
0
2
0
100
["epoch","steps"]
false
Finished
capecape
1d 7h 15m 8s
-
["/workspace/artifacts/gemma-2b:v0","google/gemma-2b"]
true
false
true
false
0.9
0.999
1.0000e-8
false
GemmaForCausalLM
false
0
false
0.01
true
false
2
0
1
-
100000
cuda:0
-
hf
pretrained=google/gemma-2b,dtype=bfloat16
false
0
false
true
1800
-
false
0
true
false
false
false
false
0
1
0
100
["epoch","steps"]
false
Failed
capecape
7d 8h 52m 46s
-
["/workspace/artifacts/gemma-7b:v0","google/gemma-7b"]
true
false
true
false
0.9
0.999
1.0000e-8
false
GemmaForCausalLM
false
0
false
0.01
true
false
2
0
1
-
100000
cuda:0
-
hf
["pretrained=/workspace/eval_harness/artifacts/gemma-7b:v0,dtype=bfloat16","pretrained=/workspace/eval_harness/artifacts/gemma-7b:v1,dtype=bfloat16","pretrained=google/gemma-7b,dtype=bfloat16"]
false
0
false
true
1800
-
false
0
true
false
false
false
false
0
1
0
100
["epoch","steps"]
false
1-3
of 3