Skip to content

insop/kd-llm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Experimental configurations and scripts

This repository contains experimental configurations and scripts for fine-tuning and knowledge distillation.

We use torchtune training framework in a separate repository. In thi repository contains torchtune configuraiton yaml files at configs/llama3_2 and runner scripts at scripts/llama3/runner.sh.

The directory structure is as follows:

-> % tree -d -L 3
.
├── bin : docker scripts
├── configs : torch tune configs
│   └── llama3_2
│       ├── ft
│       └── kd
├── docs
├── scripts
│   ├── llama3 : run scripts
│   │   ├── eval
│   │   ├── ft
│   │   └── kd
│   └── notebooks
├── sft : stsandalone fine-tuning
└── sft_script

Environment

We use the docker based environment.

  • start docker
$ scripts/docker_start.sh
  • connect to docker
$ scripts/docker_connect.sh
  • stop docker
$ scripts/docker_stop.sh

How to run

Run the scripts inside docker

# FT training
cd /workspace
/workspace# workspace/scripts/llama3/runner.sh ft/ft_train_llama3_1-8b_completion.sh  

# KD training
/workspace# workspace/scripts/llama3/runner.sh kd/train_llama3_1-8b-llama3_2-1b_sym_kld.sh

Acknoledgements

We acknowledge the following work:

About

Knowledge Distillation for Large Language Models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published