We’ve released a 100,000-sample Chain-of-Thought (CoT) dataset for fine-tuning local reasoning models.
Each sample includes explicit intermediate reasoning traces, rather than answer-only supervision. The goal is to improve reasoning consistency during supervised fine-tuning, especially for smaller local models.
We’re sharing it here to gather feedback from people working on local LLM fine-tuning and reasoning distillation.
I’d especially love feedback on:
- CoT length
- consistency of reasoning style
- whether full reasoning traces help or hurt smaller local models
Hugging Face:
https://huggingface.co/datasets/Kamisori-daijin/email-datasets-v2-100k
[link] [comments]




