Nonenman's group workspace
mlm_bipar
What makes this group special?
Tags
30_roberta-large_A100_48_384_2e-5_20-words_20epochs
Notes
pretraining roberta-large on bipar
Author
State
Finished
Start time
December 1st, 2023 3:22:21 PM
Runtime
36m 42s
Tracked hours
36m 40s
Run path
nonenman/RC_project/66ca88c5
OS
Linux-4.18.0-372.46.1.el8_6.x86_64-x86_64-with-glibc2.28
Python version
3.10.0
Command
/pfs/data5/home/as/as_as/as_nonenman/python_scripts/train_mlm.py --model_name roberta-large --batch_size 48 --seq_length 384 --num_train_epochs 20 --learning_rate 2e-5 --mask_words True --mlm_probability 0.20
System Hardware
| CPU count | 64 |
| Logical CPU count | 128 |
| GPU count | 1 |
| GPU type | NVIDIA A100 80GB PCIe |
W&B CLI Version
0.14.0
Group
mlm_biparSummary
Summary metrics are your model's outputs. Learn more
- {} 13 keys▶
- 0.9327388689631508
- 2.545281303322198
- 1,024
- 514
- 16
- 24
- 355,412,057
- 50,265
- 0.7320750010241369
- 0
- 0.7438557147979736
- 77.91145557830338
- 36.21204754027227
Artifact Outputs
This run produced these artifacts as outputs. Total: 1. Learn more
Type
Name
Consumer count
Loading...
nonenman