Urakiny's workspace
Runs
130
Name
129 visualized
State
Notes
User
Tags
Created
Runtime
Sweep
_name_or_path
activation
adafactor
adam_beta1
adam_beta2
adam_epsilon
add_cross_attention
architectures
attention_dropout
auto_find_batch_size
bf16
bf16_full_eval
chunk_size_feed_forward
dataloader_drop_last
dataloader_num_workers
dataloader_pin_memory
ddp_timeout
debug
dim
disable_tqdm
diversity_penalty
do_eval
do_predict
do_sample
do_train
dropout
early_stopping
encoder_no_repeat_ngram_size
eval_delay
evaluation_strategy
fp16
fp16_backend
fp16_full_eval
fp16_opt_level
fsdp
fsdp_config.min_num_params
fsdp_config.xla
fsdp_config.xla_fsdp_grad_ckpt
fsdp_min_num_params
full_determinism
Crashed
Fine tune model distilbert base uncased
urakiny
5m 32s
-
facebook/wav2vec2-base
-
false
0.9
0.999
1.0000e-8
false
["Wav2Vec2ForPreTraining"]
0.1
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Crashed
Fine-tuning casual language models
urakiny
3h 34m 31s
-
microsoft/phi-2
-
false
0.9
0.999
1.0000e-8
false
["PhiForCausalLM"]
0
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Crashed
Fine-tuning casual language models
urakiny
2h 13m 30s
-
microsoft/phi-2
-
false
0.9
0.999
1.0000e-8
false
["PhiForCausalLM"]
0
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Crashed
Fine tune model distilbert base uncased
urakiny
20m
-
distilgpt2
-
false
0.9
0.999
1.0000e-8
false
["GPT2LMHeadModel"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Crashed
Fine-tuning casual language models
urakiny
3h 50m 1s
-
microsoft/phi-2
-
false
0.9
0.999
1.0000e-8
false
["PhiForCausalLM"]
0
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Crashed
Fine tune model distilbert base uncased
urakiny
5m 24s
-
google/vit-base-patch16-224-in21k
-
false
0.9
0.999
1.0000e-8
false
["ViTModel"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Crashed
Fine tune model distilbert base uncased
urakiny
3m 15s
-
t5-small
-
false
0.9
0.999
1.0000e-8
false
["T5ForConditionalGeneration"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
3m 49s
-
microsoft/git-base
-
false
0.9
0.999
1.0000e-8
false
["GitForCausalLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
7m 15s
-
google/vit-base-patch16-224-in21k
-
false
0.9
0.999
1.0000e-8
false
["ViTModel"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
14m 48s
-
distilgpt2
-
false
0.9
0.999
1.0000e-8
false
["GPT2LMHeadModel"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
6m 44s
-
distilbert-base-uncased
gelu
false
0.9
0.999
1.0000e-8
false
["DistilBertForMaskedLM"]
0.1
false
false
false
0
false
0
true
1800
[]
768
false
0
true
false
false
false
0.1
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
1m 16s
-
t5-small
-
false
0.9
0.999
1.0000e-8
false
["T5ForConditionalGeneration"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
2m 7s
-
microsoft/git-base
-
false
0.9
0.999
1.0000e-8
false
["GitForCausalLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
10m 22s
-
t5-small
-
false
0.9
0.999
1.0000e-8
false
["T5ForConditionalGeneration"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine-tuning casual language models
urakiny
3m 47s
-
microsoft/phi-2
-
false
0.9
0.999
1.0000e-8
false
["PhiForCausalLM"]
0
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Finished
-
urakiny
1m 49s
-
distilgpt2
-
false
0.9
0.999
1.0000e-8
false
["GPT2LMHeadModel"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
57m 33s
-
ft-detr-with-cppe-5
-
false
0.9
0.999
1.0000e-8
false
["DetrForObjectDetection"]
0
false
false
false
0
false
0
true
1800
[]
-
false
0
false
false
false
false
0.1
false
0
0
no
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
43s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
Fine tune model distilbert base uncased
urakiny
4m 36s
-
distilbert-base-uncased
gelu
false
0.9
0.999
1.0000e-8
false
["DistilBertForTokenClassification"]
0.1
false
false
false
0
false
0
true
1800
[]
768
false
0
true
false
false
false
0.1
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
12m 45s
-
microsoft/layoutlmv3-base
-
false
0.9
0.999
1.0000e-8
false
-
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
1h 3m 21s
-
microsoft/layoutlmv3-base
-
false
0.9
0.999
1.0000e-8
false
-
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
false
auto
false
O1
[]
0
false
false
0
false
Finished
-
urakiny
1m 50s
-
distilgpt2
-
false
0.9
0.999
1.0000e-8
false
["GPT2LMHeadModel"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model bert-base-cased
urakiny
3m 52s
-
bert-base-cased
-
false
0.9
0.999
1.0000e-8
false
["BertForMaskedLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
-
urakiny
5m 35s
-
distilgpt2
-
false
0.9
0.999
1.0000e-8
false
["GPT2LMHeadModel"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
16m 32s
-
microsoft/git-base
-
false
0.9
0.999
1.0000e-8
false
["GitForCausalLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
false
false
false
false
-
false
0
0
no
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert-base-uncased
urakiny
1h 8m 36s
-
distilbert-base-uncased
gelu
false
0.9
0.999
1.0000e-8
false
["DistilBertForSequenceClassification"]
0.1
false
false
false
0
false
0
true
1800
[]
768
false
0
true
false
false
false
0.1
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
15m 28s
-
t5-small
-
false
0.9
0.999
1.0000e-8
false
["T5ForConditionalGeneration"]
-
true
false
false
0
false
0
true
1800
[]
-
false
0
false
false
false
false
-
false
0
0
no
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
9m 43s
-
google/vit-base-patch16-224-in21k
-
false
0.9
0.999
1.0000e-8
false
["ViTModel"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
1h 35m 14s
-
-
false
0.9
0.999
1.0000e-8
false
-
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
58m 6s
-
bert-base-uncased
-
false
0.9
0.999
1.0000e-8
false
["BertForMaskedLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
7m 3s
-
microsoft/git-base
-
false
0.9
0.999
1.0000e-8
false
["GitForCausalLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
11m 55s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
Fine-tuning casual language models
urakiny
3h 42m 50s
-
microsoft/phi-2
-
false
0.9
0.999
1.0000e-8
false
["PhiForCausalLM"]
0
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
15m 25s
-
t5-small
-
false
0.9
0.999
1.0000e-8
false
["T5ForConditionalGeneration"]
-
true
false
false
0
false
0
true
1800
[]
-
false
0
false
false
false
false
-
false
0
0
no
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
7m 29s
-
microsoft/git-base
-
false
0.9
0.999
1.0000e-8
false
["GitForCausalLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
1m 12s
-
nvidia/mit-b0
-
false
0.9
0.999
1.0000e-8
false
["SegformerForImageClassification"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
1m 32s
-
meta-llama/Llama-2-7b-hf
-
false
0.9
0.999
1.0000e-8
false
["LlamaForCausalLM"]
0
false
false
false
0
false
0
true
1800
[]
-
false
0
false
false
false
false
-
false
0
0
no
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
7m 8s
-
t5-small
-
false
0.9
0.999
1.0000e-8
false
["T5ForConditionalGeneration"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
7m 2s
-
dandelin/vilt-b32-mlm
-
false
0.9
0.999
1.0000e-8
false
["ViltForMaskedLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
false
false
false
false
-
false
0
0
no
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
28m 6s
-
nvidia/mit-b0
-
false
0.9
0.999
1.0000e-8
false
["SegformerForImageClassification"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
1m 36s
-
t5-small
-
false
0.9
0.999
1.0000e-8
false
["T5ForConditionalGeneration"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
2m 16s
-
microsoft/git-base
-
false
0.9
0.999
1.0000e-8
false
["GitForCausalLM"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
34m 17s
-
facebook/detr-resnet-50
-
false
0.9
0.999
1.0000e-8
false
["DetrForObjectDetection"]
0
false
false
false
0
false
0
true
1800
[]
-
false
0
false
false
false
false
0.1
false
0
0
no
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
7m 1s
-
MCG-NJU/videomae-base-finetuned-kinetics
-
false
0.9
0.999
1.0000e-8
false
["VideoMAEForVideoClassification"]
-
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
43s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
Fine tune model distilbert base uncased
urakiny
18m 35s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
Fine tune model distilbert base uncased
urakiny
42s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
Fine tune model distilbert base uncased
urakiny
42s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
Fine tune model distilbert base uncased
urakiny
6m 38s
-
facebook/wav2vec2-base
-
false
0.9
0.999
1.0000e-8
false
["Wav2Vec2ForPreTraining"]
0.1
false
false
false
0
false
0
true
1800
[]
-
false
0
true
false
false
false
-
false
0
0
steps
true
auto
false
O1
[]
0
false
false
0
false
Finished
Fine tune model distilbert base uncased
urakiny
5m 27s
-
distilbert-base-uncased
gelu
false
0.9
0.999
1.0000e-8
false
["DistilBertForMaskedLM"]
0.1
false
false
false
0
false
0
true
1800
[]
768
false
0
true
false
false
false
0.1
false
0
0
epoch
false
auto
false
O1
[]
0
false
false
0
false
1-50
of 130