Skip to content

LoRA Support #1204

@Quentin-Anthony

Description

@Quentin-Anthony

Our finetuning focus has been a bit lacking lately. Let's change that, starting with LoRA!

  • Run a finetuning baseline with a Pythia model that saturates VRAM, measure its TFLOPs/VRAM usage in wandb and link it here. Choose a small finetuning dataset for this that requires little compute to converge without being trivial.
  • Add prototype LoRA support to gpt-neox
  • Compare to baseline in step 1 and ensure TFLOP/VRAM changes make sense
  • Compare to baseline in step 1 and ensure loss is maintained

As discussed in Discord @mkerin

Metadata

Metadata

Assignees

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions