Skip to main content

Pythia 1B Trained Parameters LoRA Fine-tuning Compare

Created on May 19|Last edited on May 19

Compare Trained Parameters

Comparing LoRA fine-tuning with ShareGPT dataset, mixed English and Chinese.
  • Brown: Fine-tune after training only embed (embed_in.weight, embed_out.weight).
  • Green: Fine-tune after training embed + attention (all above + layers.n.post_attention_layernorm.weight, layers.n.post_attention_layernorm.bias, layers.n.attention.query_key_value.weight, layers.n.attention.query_key_value.bias, layers.n.attention.dense.weight, layers.n.attention.dense.bias).
  • Dark Red: Fine-tune after training all parameters.

0200m400m600m800m1train/epoch11.21.41.61.822.22.4
0.20.40.60.8train/epoch1.81.922.12.2
Run set
3