- Notifications
You must be signed in to change notification settings - Fork3
LabARSS/complexity-aware-fine-tuning
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
General-purpose Large Language Models (LLMs) are frequently fine-tuned through supervised fine-tuning (SFT) to enhance performance in specific domains. Better results can be achieved by distilling the chain-of-thought of a larger model at the cost of numerous expensive calls and a much greater amount of data.We propose a novel blueprint for efficient fine-tuning that uses reasoning only for complex data identified by entropy. Specifically, across two small open models (
Note: This is an ongoing research. If you want to reproduce the results from the EMNLP 2025 version, check outthis tag.
- DownloadCoT entropy data for MMLU to
data/out/cot_entropy - Downloadreasoning data for MMLU to
data/out/reasoning_entropy
Other datasets are included in the repo and also published on Huggingface:
uv run src/experiments/REPLACE_ME.py
@misc{goncharov2025complexityawarefinetuning, title={Complexity-aware fine-tuning}, author={Andrey Goncharov and Daniil Vyazhev and Petr Sychev and Edvard Khalafyan and Alexey Zaytsev}, year={2025}, eprint={2506.21220}, archivePrefix={arXiv}, primaryClass={cs.LG}, url={https://arxiv.org/abs/2506.21220}, }About
Resources
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Packages0
Contributors5
Uh oh!
There was an error while loading.Please reload this page.