Semantic Soft Bootstrapping (SSB)
🔗 Paper link: Arxiv preprint
🔗 Github link: Training code
Semantic Soft Bootstrapping (SSB) is an RL-free self-distillation framework that improves long-context reasoning in LLMs by training the model on its own hinted reasoning as a teacher. Rather than relying on a separate larger teacher or on-policy gradient with sparse rewards, SSB uses the same base model in two semantic roles: a hinted teacher that sees both correct and incorrect solutions and synthesizes a robust explanation, and a hint-free student that learns to reproduce this behavior from the bare question alone. Starting from a raw problem–answer dataset, we construct paired teacher–student conversations and then precompute teacher logits over the answer tokens, enabling efficient offline distillation without any human annotation or online RL loop. This is depicted as following:
Our experiments on unsloth/Qwen2.5-3B-Instruct show a gain of 10.6%, and 10% improvements in accuracy on MATH500 and AIME2024 benchmarks, compared to just GRPO based RLVR. The results are shown below:
Citation
If you find our work useful, consider citing it as:
@article{mitra2025semantic,
title={Semantic Soft Bootstrapping: Long Context Reasoning in LLMs without Reinforcement Learning},
author={Mitra, Purbesh and Ulukus, Sennur},
journal={arXiv preprint arXiv:2512.05105},
year={2025}
}