推文概览
查看 @iScienceLuvr 在 2025年5月7日 10:02 发布的这条 X/Twitter 推文。 这条内容包含 1 张图片。
Rewriting Pre-Training Data Boosts LLM Performance in Math and Code Introduces two openly licensed datasets: 1. SwallowCode (≈16.1 billion tokens) refines Python snippets from The-Stack-v2 2. SwallowMath (≈2.3 billion tokens) enhances Finemath-4+ by removing boilerplate, restoring context, and reformatting solutions into concise, step-by-step explanations







