Primary
Runs
77
Name
7 visualized
lora_target_modules
axolotl_config.lora_target_linear
axolotl_config.gradient_checkpointing
-
true
true
-
true
false
-
true
true
-
-
true
-
true
true
-
true
true
-
true
true
-
true
false
-
-
false
-
-
true
-
-
true
-
-
true
-
true
true
-
true
true
-
true
true
-
true
true
-
true
true
-
true
true
-
true
true
-
true
true
State
Notes
User
Tags
Created
Runtime
Sweep
_name_or_path
adafactor
adam_beta1
adam_beta2
adam_epsilon
adapter
add_cross_attention
architectures
attention_dropout
auto_find_batch_size
axolotl_config.adapter
axolotl_config.axolotl_config_path
axolotl_config.base_model
axolotl_config.bf16
axolotl_config.dataset_prepared_path
axolotl_config.datasets
axolotl_config.eval_table_max_new_tokens
axolotl_config.evals_per_epoch
axolotl_config.flash_attention
axolotl_config.fp16
axolotl_config.gradient_accumulation_steps
axolotl_config.group_by_length
axolotl_config.is_llama_derived_model
axolotl_config.is_mistral_derived_model
axolotl_config.learning_rate
axolotl_config.load_in_4bit
axolotl_config.load_in_8bit
axolotl_config.logging_steps
axolotl_config.lora_alpha
axolotl_config.lora_dropout
axolotl_config.lora_r
axolotl_config.loss_watchdog_patience
axolotl_config.loss_watchdog_threshold
axolotl_config.lr_scheduler
axolotl_config.max_steps
axolotl_config.micro_batch_size
axolotl_config.model_type
axolotl_config.num_epochs
axolotl_config.optimizer
axolotl_config.output_dir
axolotl_config.pad_to_sequence_len
axolotl_config.sample_packing
axolotl_config.save_strategy
Finished
-
capecape
1m 36s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
16
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
28s
-
TinyLlama/TinyLlama-1.1B-Chat-v1.0
false
0.9
0.999
1.0000e-8
-
false
["LlamaForCausalLM"]
0
false
qlora
/mixtral/tinyllama.yml
TinyLlama/TinyLlama-1.1B-Chat-v1.0
true
-
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
1
false
true
-
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
1
LlamaForCausalLM
1
adamw_bnb_8bit
./model-out
true
true
steps
Finished
-
capecape
1m 22s
-
TinyLlama/TinyLlama-1.1B-Chat-v1.0
false
0.9
0.999
1.0000e-8
-
false
["LlamaForCausalLM"]
0
false
qlora
/mixtral/tinyllama.yml
TinyLlama/TinyLlama-1.1B-Chat-v1.0
true
-
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
4
false
true
-
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
1
LlamaForCausalLM
1
adamw_bnb_8bit
./model-out
true
true
steps
Finished
-
capecape
1m 21s
-
TinyLlama/TinyLlama-1.1B-Chat-v1.0
false
0.9
0.999
1.0000e-8
-
false
["LlamaForCausalLM"]
0
false
-
/mixtral/tinyllama.yml
TinyLlama/TinyLlama-1.1B-Chat-v1.0
true
-
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
4
false
true
-
0.0002
false
false
1
-
-
-
3
5
cosine
50
1
LlamaForCausalLM
1
adamw_bnb_8bit
./model-out
true
true
steps
Finished
-
capecape
19m 33s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
16
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
19m 31s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"type":"alpaca","path":"vicgalle/alpaca-gpt4"}]
128
-
true
false
16
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
19m 26s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"type":"alpaca","path":"vicgalle/alpaca-gpt4"}]
128
-
true
false
16
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
13m 33s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
16
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
11m 40s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
16
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
12m 58s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
4
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
12m 58s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
4
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
13m 24s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
4
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
head
mistral-qlora
target-linear
15m 11s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
4
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
mistral-qlora
target-linear
15m 4s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
-
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
4
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
1m 45s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"type":"alpaca","path":"vicgalle/alpaca-gpt4"}]
128
4
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
50
4
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
15m 13s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"type":"alpaca","path":"vicgalle/alpaca-gpt4"}]
128
4
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
200
2
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
22m 56s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
./dataset_prepare
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
4
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
400
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
2m 23s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
last_run_prepared
[{"path":"vicgalle/alpaca-gpt4","type":"alpaca"}]
128
4
true
false
4
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
400
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
3m 43s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
last_run_prepared
[{"path":"mhenrichsen/alpaca_2k_test","type":"alpaca"}]
128
4
true
false
1
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
10
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
Finished
-
capecape
4m 3s
-
mistralai/Mistral-7B-v0.1
false
0.9
0.999
1.0000e-8
-
false
["MistralForCausalLM"]
0
false
qlora
/mixtral/mistral.yml
mistralai/Mistral-7B-v0.1
true
last_run_prepared
[{"type":"alpaca","path":"mhenrichsen/alpaca_2k_test"}]
128
4
true
false
1
false
-
true
0.0002
true
false
1
16
0.05
32
3
5
cosine
10
1
MistralForCausalLM
1
adamw_bnb_8bit
./qlora-out
true
true
steps
1-20
of 55