NLP [2026-1] 백승우 - Self-Improving Pretraining:using post-trained models to pretrain better models BaekDaBang 2026. 2. 4. 13:47 Self-Improving Pretraining: using post-trained models to pretrain better models Ensuring safety, factuality and overall quality in the generations of large language models is a critical challenge, especially as these models are increasingly deployed in real-world applications. The prevailing approach to addressing these issues involve arxiv.org