Dchanda's group workspace
Group: k5nu8k69390a-Baseline
State
Notes
User
Tags
Created
Runtime
Sweep
T_max
device
epochs
group
hash_name
hidden_state
learning_rate
margin
max_length
min_lr
model_name
n_accumulate
n_fold
num_classes
scheduler
seed
tokenizer
train_batch_size
valid_batch_size
weight_decay
Best Loss
Train Loss
Valid Loss
Finished
-
dchanda
2h 54m 33s
-
500
cuda:0
3
k5nu8k69390a-Baseline
k5nu8k69390a
-
0.0001
0.5
128
0.000001
roberta-base
1
5
1
CosineAnnealingLR
2021
PreTrainedTokenizerFast(name_or_path='roberta-base', vocab_size=50265, model_max_len=512, is_fast=True, padding_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': AddedToken("<mask>", rstrip=False, lstrip=True, single_word=False, normalized=False)})
32
64
0.000001
0.34013
0.33507
0.35942
Finished
-
dchanda
k5nu8k69390a
margin-loss
roberta-base
34m 44s
-
500
cuda:0
3
k5nu8k69390a-Baseline
k5nu8k69390a
-
0.0001
0.5
128
0.000001
roberta-base
1
5
1
CosineAnnealingLR
2021
PreTrainedTokenizerFast(name_or_path='roberta-base', vocab_size=50265, model_max_len=512, is_fast=True, padding_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': AddedToken("<mask>", rstrip=False, lstrip=True, single_word=False, normalized=False)})
32
64
0.000001
0.34013
0.3468
0.34972
Finished
-
dchanda
k5nu8k69390a
margin-loss
roberta-base
34m 49s
-
500
cuda:0
3
k5nu8k69390a-Baseline
k5nu8k69390a
-
0.0001
0.5
128
0.000001
roberta-base
1
5
1
CosineAnnealingLR
2021
PreTrainedTokenizerFast(name_or_path='roberta-base', vocab_size=50265, model_max_len=512, is_fast=True, padding_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': AddedToken("<mask>", rstrip=False, lstrip=True, single_word=False, normalized=False)})
32
64
0.000001
0.33951
0.32539
0.34738
Finished
-
dchanda
k5nu8k69390a
margin-loss
roberta-base
34m 48s
-
500
cuda:0
3
k5nu8k69390a-Baseline
k5nu8k69390a
-
0.0001
0.5
128
0.000001
roberta-base
1
5
1
CosineAnnealingLR
2021
PreTrainedTokenizerFast(name_or_path='roberta-base', vocab_size=50265, model_max_len=512, is_fast=True, padding_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': AddedToken("<mask>", rstrip=False, lstrip=True, single_word=False, normalized=False)})
32
64
0.000001
0.34186
0.32772
0.34267
Finished
-
dchanda
k5nu8k69390a
margin-loss
roberta-base
34m 45s
-
500
cuda:0
3
k5nu8k69390a-Baseline
k5nu8k69390a
-
0.0001
0.5
128
0.000001
roberta-base
1
5
1
CosineAnnealingLR
2021
PreTrainedTokenizerFast(name_or_path='roberta-base', vocab_size=50265, model_max_len=512, is_fast=True, padding_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': AddedToken("<mask>", rstrip=False, lstrip=True, single_word=False, normalized=False)})
32
64
0.000001
0.33439
0.32449
0.3433
Finished
-
dchanda
k5nu8k69390a
margin-loss
roberta-base
35m 15s
-
500
cuda:0
3
k5nu8k69390a-Baseline
k5nu8k69390a
-
0.0001
0.5
128
0.000001
roberta-base
1
5
1
CosineAnnealingLR
2021
PreTrainedTokenizerFast(name_or_path='roberta-base', vocab_size=50265, model_max_len=512, is_fast=True, padding_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': AddedToken("<mask>", rstrip=False, lstrip=True, single_word=False, normalized=False)})
32
64
0.000001
0.34474
0.35094
0.41403
1-1
of 1