Chinese artificial intelligence firm Moonshot AI has introduced an open-source middleware solution named Checkpoint Engine, tailored explicitly for large language model (LLM) inference engines. This middleware boasts in-place hot-updating capabilities for model weights, rendering it highly suitable for reinforcement learning environments. It is capable of synchronizing the weights of the Kimi-K2 model—which boasts a staggering 1 trillion parameters—within a mere 20 seconds, all while supporting parallel operations across thousands of GPUs. Checkpoint Engine is intricately integrated with the vLLM framework and is made available under the Apache 2.0 license. Furthermore, it is designed to be extensible and compatible with other frameworks, such as SGLang, thereby significantly enhancing computational efficiency and accelerating development cycles.