Skip to content

Commit 6c298f2

Browse files
committed
fix model names in mocks for tokenizer
1 parent 4674897 commit 6c298f2

File tree

5 files changed

+40
-27
lines changed

5 files changed

+40
-27
lines changed

src/components/ChatInput.tsx

Lines changed: 16 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -146,7 +146,7 @@ export const ChatInput: React.FC<ChatInputProps> = ({
146146
const deferredModel = useDeferredValue(preferredModel);
147147
const deferredInput = useDeferredValue(input);
148148
const tokenCountPromise = useMemo(() => {
149-
if (!deferredModel || deferredInput.trim().length === 0) {
149+
if (!deferredModel || deferredInput.trim().length === 0 || deferredInput.startsWith("/")) {
150150
return Promise.resolve(0);
151151
}
152152
return getTokenCountPromise(deferredModel, deferredInput);
@@ -860,19 +860,21 @@ export const ChatInput: React.FC<ChatInputProps> = ({
860860
<div className="max-@[500px]:hidden flex items-center" data-component="Context1MGroup">
861861
<Context1MCheckbox modelString={preferredModel} />
862862
</div>
863-
{preferredModel && hasTypedText && (
864-
<Suspense
865-
fallback={
866-
<div
867-
className="text-muted flex items-center gap-1 text-xs"
868-
data-component="TokenEstimate"
869-
>
870-
<span>Calculating tokens…</span>
871-
</div>
872-
}
873-
>
874-
<TokenCountDisplay reader={tokenCountReader} />
875-
</Suspense>
863+
{preferredModel && (
864+
<div className={hasTypedText ? "block" : "hidden"}>
865+
<Suspense
866+
fallback={
867+
<div
868+
className="text-muted flex items-center gap-1 text-xs"
869+
data-component="TokenEstimate"
870+
>
871+
<span>Calculating tokens…</span>
872+
</div>
873+
}
874+
>
875+
<TokenCountDisplay reader={tokenCountReader} />
876+
</Suspense>
877+
</div>
876878
)}
877879
<div className="max-@[700px]:hidden ml-auto flex items-center gap-1.5">
878880
<div

src/services/mock/scenarios/permissionModes.ts

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ const executePlanTurn: ScenarioTurn = {
7474
kind: "stream-start",
7575
delay: 0,
7676
messageId: "msg-exec-refactor",
77-
model: "mock:executor",
77+
model: "openai:gpt-5",
7878
},
7979
{
8080
kind: "tool-start",
@@ -118,7 +118,7 @@ const executePlanTurn: ScenarioTurn = {
118118
kind: "stream-end",
119119
delay: STREAM_BASE_DELAY * 3,
120120
metadata: {
121-
model: "mock:executor",
121+
model: "openai:gpt-5",
122122
inputTokens: 220,
123123
outputTokens: 110,
124124
systemMessageTokens: 18,

src/services/mock/scenarios/review.ts

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,7 @@ const openOnboardingDocTurn: ScenarioTurn = {
8686
assistant: {
8787
messageId: "msg-exec-1",
8888
events: [
89-
{ kind: "stream-start", delay: 0, messageId: "msg-exec-1", model: "mock:executor" },
89+
{ kind: "stream-start", delay: 0, messageId: "msg-exec-1", model: "openai:gpt-5" },
9090
{
9191
kind: "tool-start",
9292
delay: STREAM_BASE_DELAY,
@@ -114,7 +114,7 @@ const showOnboardingDocTurn: ScenarioTurn = {
114114
assistant: {
115115
messageId: "msg-exec-2",
116116
events: [
117-
{ kind: "stream-start", delay: 0, messageId: "msg-exec-2", model: "mock:executor" },
117+
{ kind: "stream-start", delay: 0, messageId: "msg-exec-2", model: "openai:gpt-5" },
118118
{
119119
kind: "tool-start",
120120
delay: STREAM_BASE_DELAY,
@@ -153,7 +153,7 @@ const showOnboardingDocTurn: ScenarioTurn = {
153153
kind: "stream-end",
154154
delay: STREAM_BASE_DELAY * 3,
155155
metadata: {
156-
model: "mock:executor",
156+
model: "openai:gpt-5",
157157
inputTokens: 96,
158158
outputTokens: 142,
159159
systemMessageTokens: 32,

src/services/mock/scenarios/toolFlows.ts

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -19,7 +19,7 @@ const fileReadTurn: ScenarioTurn = {
1919
assistant: {
2020
messageId: "msg-tool-file-read",
2121
events: [
22-
{ kind: "stream-start", delay: 0, messageId: "msg-tool-file-read", model: "mock:executor" },
22+
{ kind: "stream-start", delay: 0, messageId: "msg-tool-file-read", model: "openai:gpt-5" },
2323
{
2424
kind: "tool-start",
2525
delay: STREAM_BASE_DELAY,
@@ -55,7 +55,7 @@ const fileReadTurn: ScenarioTurn = {
5555
kind: "stream-end",
5656
delay: STREAM_BASE_DELAY * 3,
5757
metadata: {
58-
model: "mock:executor",
58+
model: "openai:gpt-5",
5959
inputTokens: 92,
6060
outputTokens: 64,
6161
systemMessageTokens: 18,
@@ -78,7 +78,7 @@ const listDirectoryTurn: ScenarioTurn = {
7878
assistant: {
7979
messageId: "msg-tool-bash-ls",
8080
events: [
81-
{ kind: "stream-start", delay: 0, messageId: "msg-tool-bash-ls", model: "mock:executor" },
81+
{ kind: "stream-start", delay: 0, messageId: "msg-tool-bash-ls", model: "openai:gpt-5" },
8282
{
8383
kind: "tool-start",
8484
delay: STREAM_BASE_DELAY,
@@ -122,7 +122,7 @@ const listDirectoryTurn: ScenarioTurn = {
122122
kind: "stream-end",
123123
delay: STREAM_BASE_DELAY * 3,
124124
metadata: {
125-
model: "mock:executor",
125+
model: "openai:gpt-5",
126126
inputTokens: 74,
127127
outputTokens: 58,
128128
systemMessageTokens: 16,
@@ -151,7 +151,7 @@ const createTestFileTurn: ScenarioTurn = {
151151
kind: "stream-start",
152152
delay: 0,
153153
messageId: "msg-tool-create-test-file",
154-
model: "mock:executor",
154+
model: "openai:gpt-5",
155155
},
156156
{
157157
kind: "tool-start",
@@ -181,7 +181,7 @@ const createTestFileTurn: ScenarioTurn = {
181181
kind: "stream-end",
182182
delay: STREAM_BASE_DELAY * 3,
183183
metadata: {
184-
model: "mock:executor",
184+
model: "openai:gpt-5",
185185
inputTokens: 80,
186186
outputTokens: 40,
187187
systemMessageTokens: 12,
@@ -205,7 +205,7 @@ const readTestFileTurn: ScenarioTurn = {
205205
kind: "stream-start",
206206
delay: 0,
207207
messageId: "msg-tool-read-test-file",
208-
model: "mock:executor",
208+
model: "openai:gpt-5",
209209
},
210210
{
211211
kind: "tool-start",
@@ -242,7 +242,7 @@ const readTestFileTurn: ScenarioTurn = {
242242
kind: "stream-end",
243243
delay: STREAM_BASE_DELAY * 3,
244244
metadata: {
245-
model: "mock:executor",
245+
model: "openai:gpt-5",
246246
inputTokens: 76,
247247
outputTokens: 52,
248248
systemMessageTokens: 12,

src/styles/globals.css

Lines changed: 11 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -831,6 +831,17 @@ pre code {
831831
}
832832
}
833833

834+
@keyframes toastSlideOut {
835+
from {
836+
transform: translateX(0);
837+
opacity: 1;
838+
}
839+
to {
840+
transform: translateX(20px);
841+
opacity: 0;
842+
}
843+
}
844+
834845
@keyframes ellipsis {
835846
0% {
836847
content: "";

0 commit comments

Comments
 (0)