Skip to content

[Question] Guidance on training steps vs. dataset size for LoRA fine-tuning #61

@fyting

Description

@fyting

Hi there,

First of all, thanks for the great work on this repository!

I have a question regarding the relationship between the dataset size and the number of training steps during LoRA fine-tuning. I noticed that in the provided example configs, steps is often set to fixed values like 1500 or 2000.

Is there a rule of thumb or empirical formula for adjusting the steps if my dataset size varies significantly? For example, should I aim for a specific number of epochs, or is the fixed step count generally sufficient regardless of the data scale?

Any advice or experience would be greatly appreciated. Thanks!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions