Make sure to install the requirements for these examples:
pip install -r requirements-examples.txtExamples for training control vectors from a model's hidden states that can be used to influence the behavior and generated output during inference.
LoRA training Mistral-7B-Instruct-v0.2 with the Nvidia HelpSteer dataset.
Run train.sh in the helpsteer directory to download the dataset & model from HuggingFace and start the LoRA training. You can customize the training configuration by editing config.yml.
DPO training Qwen1.5-7B-Chat with the DPO Mix 7K dataset. The training consists of a supervised fine tuning (SFT) followed by direct preference optimization (DPO).
Run train.sh in the dpo-mix-7k directory to download the dataset & model from HuggingFace and start the training. You can customize the training configuration by editing the config files sft.yml and dpo.yml.
Implementation of the "Massive Multitask Language Understanding" benchmark using the MMLU dataset.
Run mmlu.py with the model you would like to evaluate.
Implementation of the MMLU-Pro benchmark using the MMLU-Pro dataset.
Run mmlu-pro.py with the model you would like to evaluate.
Calculating perplexity scores for a sample dataset of entry paragraphs from Wikipedia articles.
Run perplexity.py with the model you would like to evaluate. Add quantization options to evaluate perplexity with quantized models.