Skip to content

Commit 2a84a24

Browse files
authored
Update 2025-11-10-bitwise-consistent-train-inference.md
Signed-off-by: Bram Wasti <[email protected]>
1 parent 08c3cf6 commit 2a84a24

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

_posts/2025-11-10-bitwise-consistent-train-inference.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ For the RL demo, we wrote a generic reinforcement learning script using GSM8K an
2727

2828
## What’s Next
2929

30-
We will continue to push forward on bitwise consistent training and inference. To follow this work, please see the linked RFC: [#28326](https://github.com/vllm-project/vllm/issues/28326). More specifically, we will focus on the following directions:
30+
We will continue to push forward on bitwise consistent training and inference. To follow this work, please see the linked RFCs: [#28326](https://github.com/vllm-project/vllm/issues/28326) and [#27433](https://github.com/vllm-project/vllm/issues/27433). More specifically, we will focus on the following directions:
3131

3232
**Unified model definition.** Although we have demonstrated the bitwise equivalent training and inference results, there are still two copies of the model code, one for training and one for inference. This is easy for our first integration but fragile for long-term maintenance: any slight change to each of the model code will break the equivalence between training and inference and lead to numerical mismatches. Having a shared model code for both training and inference frameworks will eliminate the possibility of introducing accidental human errors and make the bitwise matching property easier to maintain.
3333

0 commit comments

Comments
 (0)