NLP

[2026-1] 백승우 - Self-Improving Pretraining:using post-trained models to pretrain better models

BaekDaBang 2026. 2. 4. 13:47
 

Self-Improving Pretraining: using post-trained models to pretrain better models

Ensuring safety, factuality and overall quality in the generations of large language models is a critical challenge, especially as these models are increasingly deployed in real-world applications. The prevailing approach to addressing these issues involve

arxiv.org