distily
1.0.0
pip install -U "git+https://github.com/lapp0/distily.git"
Distily allows you to distill a model with
Minimal Example: distily_gpt2
Command to create a distilled gpt2 with only 6 layers:
python3 -m distily.run
--teacher_model_name_or_path gpt2
--output_dir distily_gpt2
--hub_model_id "distily/distily_gpt2"
--push_to_hub True
--student_model_config {"n_layers": 6}
--student_model_as_bitnet True
The Resulting distily_gpt2 Model has (TODO: explain metrics).
For more examples, review the Examples documentation.
To push to hub, you must prepare your hub token
HF_WRITE=<your hub token> python3 -c "from huggingface_hub.hf_api import HfFolder; HfFolder.save_token('${HF_WRITE}')"
TODO: commit the linked docs once complete
Using Distily
Available Models
Contributing
num_hidden_layers by implementing layer mappers.intermediate_size, hidden_act) by employing projectors.num_attention_heads and num_key_value_heads.