Skip to main content

Nonenman's group workspace

mlm_bipar

What makes this group special?
Tags

30_roberta-large_A100_48_384_2e-5_20-words_20epochs

Notes

pretraining roberta-large on bipar

Author
State
Finished
Start time
December 1st, 2023 3:22:21 PM
Runtime
36m 42s
Tracked hours
36m 40s
Run path
nonenman/RC_project/66ca88c5
OS
Linux-4.18.0-372.46.1.el8_6.x86_64-x86_64-with-glibc2.28
Python version
3.10.0
Command
/pfs/data5/home/as/as_as/as_nonenman/python_scripts/train_mlm.py --model_name roberta-large --batch_size 48 --seq_length 384 --num_train_epochs 20 --learning_rate 2e-5 --mask_words True --mlm_probability 0.20
System Hardware
CPU count64
Logical CPU count 128
GPU count1
GPU typeNVIDIA A100 80GB PCIe
W&B CLI Version
0.14.0
Config

Config parameters are your model's inputs. Learn more

  • {} 7 keys
    • 48
    • 0.00002
    • 0.2
    • 20
    • 384
    • 0.06
    • 0.01
Summary

Summary metrics are your model's outputs. Learn more

  • {} 13 keys
    • 0.9327388689631508
    • 2.545281303322198
    • 1,024
    • 514
    • 16
    • 24
    • 355,412,057
    • 50,265
    • 0.7320750010241369
    • 0
    • 0.7438557147979736
    • 77.91145557830338
    • 36.21204754027227
Artifact Outputs

This run produced these artifacts as outputs. Total: 1. Learn more

Loading...