Skip to content

Latest commit

 

History

History

README.md

Direct Preference Knowledge Distillation for Large Language Models

Environment

conda create -n dpkd python=3.11
conda activate dpkd

and

bash install.sh

Run Distillation

Train runner:

bash scripts/dpkd-gpt2_base_runner.sh  PATH_TO_DPKD 

Evaluation

Evaluation runner:

bash scripts/dpkd-gpt2_base_evaluate.sh  PATH_TO_DPKD 

Citation

@article{dpkd,
  title={Direct Preference Knowledge Distillation for Large Language Models},
  author={Yixing Li and Yuxian Gu and Li Dong and Dequan Wang and Yu Cheng and Furu Wei},
  journal={arXiv preprint arXiv:2406.19774},
  year={2024}
}