Hey everyone,

I came across a post recently where someone found it hard to find simple scripts to fine-tune LLMs with their data. So I put together a repo where you can just type out your requirements in a config.yaml file and the training happens flawlessly based on that.

Here’s the repo - LLM-Trainer

It is still a wip so lemme know if guys want some other features added to this.

TIA.

    • Dry_Long3157@alien.topOPB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Hey, you could just download the config file and lora_train.py file and run it as I’ve explained in the readme!

      To simplify it further, open both the file in any editor and load up the same environment you use for oobabooga. Then make all the changes based on your req in the lora_config.yaml file. Once you’re done just run “python lora_train.py”.

      If you need further help, feel free to ask!

  • uhuge@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I assume auth_token is for storing the merged model in HF? Seems worth noting/clarifying.

    I’ll get back with more feedback when I get to test it.+)