Andberg9's workspace
Runs
11
Name
11 visualized
State
Notes
User
Tags
Created
Runtime
Sweep
augmentations.augmentations
augmentations.augmentations_per_sample
general_config
loss_config
method
metric.goal
metric.name
name
num_classes
optimizer_config
parameters.peft_scheduling.schedule.0.peft_methods
parameters.peft_scheduling.schedule.0.start_epochs
parameters.peft_scheduling.schedule.1.peft_methods
parameters.peft_scheduling.schedule.1.start_epochs
parameters.peft_scheduling.schedule.2.peft_methods
parameters.peft_scheduling.schedule.2.start_epochs
parameters.peft_scheduling.schedule.3.peft_methods
parameters.peft_scheduling.schedule.3.start_epochs
peft_config
peft_scheduling.schedule.0.peft_method
peft_scheduling.schedule.0.start_epoch
peft_scheduling.schedule.1.peft_method
peft_scheduling.schedule.1.start_epoch
peft_scheduling.schedule.2.peft_method
peft_scheduling.schedule.2.start_epoch
peft_scheduling.schedule.3.peft_method
peft_scheduling.schedule.3.start_epoch
peft_scheduling_config
project
best_val_accuracy
epoch
inference_acc
inference_accuracy
inference_f1
inference_precision
inference_recall
lr-Adam
lr-AdamW
model/memory_footprint_mb
model/total_parameters
model/trainable_parameters
model/trainable_percentage
peft_method_epoch
total_train_time
Finished
andberg9
14s
["time_mask","time_stretch","sin_distortion"]
0
model_type='resnet18' save_dataloader=False batch_size=16 seed=42 num_cuda_workers=6 pinned_memory=True epochs=20 save_model=False from_scratch=False test_size=0.0 inference_size=0.0 val_size=0.2 sweep_count=200 accumulation_steps=1 patience=5 use_wandb=True use_sweep=False torch_viz=False use_kfold=False k_folds=5 adapter_type='none-full' compute_fisher=True fisher_mc_samples=100 save_fim_heatmap=True save_fim_epochs=True early_stopping=True checkpointing=True monitor='val_acc' mode='max' save_top_k=1 test_during_training=True test_during_training_freq=1 distributed_training=False num_gpus=1 strategy='ddp'
type='cross_entropy' label_smoothing=0.0 class_weights=None focal_alpha=None focal_gamma=2.0
-
-
-
-
50
optimizer_type='adam' adam=AdamConfig(lr=0.0001, betas=(0.99, 0.999), eps=1e-08, weight_decay=0.0, amsgrad=False) adamw=AdamWConfig(lr=0.0008, betas=(0.99, 0.999), eps=1e-08, weight_decay=0.02, amsgrad=False) adamspd=AdamSPDConfig(lr=0.0001, betas=(0.9, 0.999), eps=1e-08, weight_decay=0.01, amsgrad=False) warmup=WarmupConfig(enabled=True, warmup_steps=100, warmup_start_lr=1e-06, warmup_method='linear') scheduler_type='cosine_annealing_lr' reduce_lr_on_plateau=ReduceLROnPlateauConfig(mode='max', factor=0.85, patience=3, threshold=0.0001, threshold_mode='rel', cooldown=0, min_lr=0.0, eps=1e-08) step_lr=StepLRConfig(step_size=30, gamma=0.1) cosine_annealing_lr=CosineAnnealingLRConfig(T_max=50, eta_min=0.0) gradient_clipping_enabled=True gradient_clip_val=1.0 gradient_clip_algorithm='norm'
-
-
-
-
-
-
-
-
adapter_type='none-full' task_type='SEQ_CLS'
none-full
0
ssf
3
lorac
6
batchnorm
9
enabled=False model_name=None schedule=[PEFTScheduleStep(start_epoch=0, peft_method='none-full', merge_previous=True), PEFTScheduleStep(start_epoch=3, peft_method='ssf', merge_previous=True), PEFTScheduleStep(start_epoch=6, peft_method='lorac', merge_previous=True), PEFTScheduleStep(start_epoch=9, peft_method='batchnorm', merge_previous=True)] auto_merge=True
-
-
1
-
-
-
-
-
0.00000397
-
42.70893
11195890
11195890
100
-
-
Failed
andberg9
9s
["time_mask","time_stretch","sin_distortion"]
0
model_type='ast' save_dataloader=False batch_size=16 seed=42 num_cuda_workers=6 pinned_memory=True epochs=20 save_model=False from_scratch=False test_size=0.0 inference_size=0.0 val_size=0.2 sweep_count=200 accumulation_steps=1 patience=5 use_wandb=True use_sweep=False torch_viz=False use_kfold=False k_folds=5 adapter_type='none-full' compute_fisher=True fisher_mc_samples=100 save_fim_heatmap=True save_fim_epochs=True early_stopping=True checkpointing=True monitor='val_acc' mode='max' save_top_k=1 test_during_training=True test_during_training_freq=1 distributed_training=False num_gpus=1 strategy='ddp'
type='cross_entropy' label_smoothing=0.0 class_weights=None focal_alpha=None focal_gamma=2.0
-
-
-
-
50
optimizer_type='adam' adam=AdamConfig(lr=0.0001, betas=(0.99, 0.999), eps=1e-08, weight_decay=0.0, amsgrad=False) adamw=AdamWConfig(lr=0.0008, betas=(0.99, 0.999), eps=1e-08, weight_decay=0.02, amsgrad=False) adamspd=AdamSPDConfig(lr=0.0001, betas=(0.9, 0.999), eps=1e-08, weight_decay=0.01, amsgrad=False) warmup=WarmupConfig(enabled=True, warmup_steps=100, warmup_start_lr=1e-06, warmup_method='linear') scheduler_type='cosine_annealing_lr' reduce_lr_on_plateau=ReduceLROnPlateauConfig(mode='max', factor=0.85, patience=3, threshold=0.0001, threshold_mode='rel', cooldown=0, min_lr=0.0, eps=1e-08) step_lr=StepLRConfig(step_size=30, gamma=0.1) cosine_annealing_lr=CosineAnnealingLRConfig(T_max=50, eta_min=0.0) gradient_clipping_enabled=True gradient_clip_val=1.0 gradient_clip_algorithm='norm'
-
-
-
-
-
-
-
-
adapter_type='none-full' task_type='SEQ_CLS'
oft
0
layernorm
3
layernorm
6
bitfit
9
enabled=False model_name=None schedule=[PEFTScheduleStep(start_epoch=0, peft_method='oft', merge_previous=True), PEFTScheduleStep(start_epoch=3, peft_method='layernorm', merge_previous=True), PEFTScheduleStep(start_epoch=6, peft_method='layernorm', merge_previous=True), PEFTScheduleStep(start_epoch=9, peft_method='bitfit', merge_previous=True)] auto_merge=True
-
-
-
-
-
-
-
-
0.00000199
-
328.93085
86227250
86227250
100
-
-
Failed
andberg9
2s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
layernorm
0
layernorm
3
layernorm
6
ia3
9
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Failed
andberg9
2s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
lora
0
none-full
3
none-full
6
ssf
9
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Failed
andberg9
1s
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
oft
0
bitfit
3
none-full
6
bitfit
9
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
andberg9
2m 44s
["time_mask","time_stretch","sin_distortion"]
2
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
bitfit
0
none-full
3
hra
6
ssf
9
-
-
-
-
-
-
-
-
-
-
-
328.93085
86227250
142898
0.16572
-
-
Finished
andberg9
15s
["time_mask","time_stretch","sin_distortion"]
2
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
lorac
0
lorac
3
none-full
6
none-full
9
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
andberg9
4s
["time_mask","time_stretch","sin_distortion"]
2
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
none-classifier
0
none-classifier
3
none-classifier
6
lorac
9
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
andberg9
16s
["time_mask","time_stretch","sin_distortion"]
2
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
lora
0
adalora
3
adalora
6
ssf
9
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Finished
andberg9
3s
["time_mask","time_stretch","sin_distortion"]
2
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
layernorm
0
adalora
3
lora
6
hra
9
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
Failed
andberg9
1m 8s
["time_mask","time_stretch","sin_distortion"]
2
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
layernorm
0
none-full
3
lora
6
adalora
9
-
-
-
-
-
-
-
-
-
-
-
328.93085
86227250
142898
0.16572
-
-
1-11
of 11