@@ -22,7 +22,6 @@ import (
2222 "fmt"
2323 "time"
2424
25- "github.com/google/uuid"
2625 "github.com/llm-d/llm-d-inference-sim/pkg/common"
2726 openaiserverapi "github.com/llm-d/llm-d-inference-sim/pkg/openai-server-api"
2827 "github.com/valyala/fasthttp"
@@ -154,7 +153,7 @@ func (s *VllmSimulator) sendTokenChunks(context *streamingContext, w *bufio.Writ
154153// supports both modes (text and chat)
155154func (s * VllmSimulator ) createUsageChunk (context * streamingContext , usageData * openaiserverapi.Usage ) openaiserverapi.CompletionRespChunk {
156155 baseChunk := openaiserverapi.BaseCompletionResponse {
157- ID : chatComplIDPrefix + uuid . NewString (),
156+ ID : chatComplIDPrefix + common . GenerateUUIDString (),
158157 Created : context .creationTime ,
159158 Model : context .model ,
160159 Usage : usageData ,
@@ -179,7 +178,7 @@ func (s *VllmSimulator) createUsageChunk(context *streamingContext, usageData *o
179178func (s * VllmSimulator ) createTextCompletionChunk (context * streamingContext , token string , finishReason * string ) openaiserverapi.CompletionRespChunk {
180179 return & openaiserverapi.TextCompletionResponse {
181180 BaseCompletionResponse : openaiserverapi.BaseCompletionResponse {
182- ID : chatComplIDPrefix + uuid . NewString (),
181+ ID : chatComplIDPrefix + common . GenerateUUIDString (),
183182 Created : context .creationTime ,
184183 Model : context .model ,
185184 Object : textCompletionObject ,
@@ -199,7 +198,7 @@ func (s *VllmSimulator) createChatCompletionChunk(context *streamingContext, tok
199198 role string , finishReason * string ) openaiserverapi.CompletionRespChunk {
200199 chunk := openaiserverapi.ChatCompletionRespChunk {
201200 BaseCompletionResponse : openaiserverapi.BaseCompletionResponse {
202- ID : chatComplIDPrefix + uuid . NewString (),
201+ ID : chatComplIDPrefix + common . GenerateUUIDString (),
203202 Created : context .creationTime ,
204203 Model : context .model ,
205204 Object : chatCompletionChunkObject ,
0 commit comments