Skip to main content

BigBird base NER fine-tuning results

A comparison of training results for the Feedback Prize competition
Created on December 19|Last edited on December 21


Training comparison

parameterbb-basebb-base-idpt
model namegoogle/bigbird-roberta-basegoogle/bigbird-roberta-base
in-domain pre-trainingNoYes
learning rate2e-52e-5
weight decay0.00950.0095
train batch size44
gradient accumulation steps88
fp16TrueTrue
seed1818



10203040Step0.650.70.750.80.850.9
10203040Step00.10.20.30.40.5
010203040Step0.511.522.5