Skip to content

Commit 51b200d

Browse files
committed
fix type errors
1 parent 410d410 commit 51b200d

File tree

4 files changed

+9
-135
lines changed

4 files changed

+9
-135
lines changed

examples/tool-calls-beta-zod.ts

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -100,7 +100,6 @@ async function main() {
100100
console.log();
101101
}
102102
}
103-
console.log(JSON.stringify(runner.params.messages, null, 2));
104103
}
105104

106105
main();

src/lib/beta/BetaToolRunner.ts

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -246,7 +246,7 @@ export class BetaToolRunner<Stream extends boolean> {
246246
* }
247247
*/
248248
async generateToolResponse() {
249-
// The most recent message from the assistant. This prev had this.params.messages.at(-1) but I think that's wrong.
249+
// The most recent message from the assistant. TODO: do we want || this.params.messages.at(-1)?
250250
const message = await this.#message;
251251
if (!message) {
252252
return null;

src/lib/transform-json-schema.ts

Lines changed: 0 additions & 124 deletions
This file was deleted.

tests/lib/tools/ToolRunner.test.ts

Lines changed: 8 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,16 @@
11
import OpenAI from 'openai';
22
import { mockFetch } from '../../utils/mock-fetch';
3-
import { BetaRunnableTool } from 'openai/lib/beta/BetaRunnableTool';
4-
import {
3+
import type { BetaRunnableTool } from 'openai/lib/beta/BetaRunnableTool';
4+
import type {
55
ChatCompletion,
66
ChatCompletionChunk,
77
ChatCompletionMessage,
88
ChatCompletionMessageFunctionToolCall,
99
ChatCompletionMessageToolCall,
1010
ChatCompletionToolMessageParam,
1111
} from 'openai/resources';
12-
import { Fetch } from 'openai/internal/builtin-types';
12+
import type { Fetch } from 'openai/internal/builtin-types';
13+
import type { BetaToolRunnerParams } from 'openai/lib/beta/BetaToolRunner';
1314

1415
const weatherTool: BetaRunnableTool<{ location: string }> = {
1516
type: 'function',
@@ -286,13 +287,11 @@ interface SetupTestResult<Stream extends boolean> {
286287
handleAssistantMessageStream: (messageContentOrToolCalls?: ToolCallsOrMessage) => ChatCompletion;
287288
}
288289

289-
type ToolRunnerParams = Parameters<typeof OpenAI.Beta.Chat.Completions.prototype.toolRunner>[0];
290-
291290
type ToolCallsOrMessage = ChatCompletionMessageToolCall[] | ChatCompletionMessage;
292291

293-
function setupTest(params?: Partial<ToolRunnerParams> & { stream?: false }): SetupTestResult<false>;
294-
function setupTest(params: Partial<ToolRunnerParams> & { stream: true }): SetupTestResult<true>;
295-
function setupTest(params: Partial<ToolRunnerParams> = {}): SetupTestResult<boolean> {
292+
function setupTest(params?: Partial<BetaToolRunnerParams> & { stream?: false }): SetupTestResult<false>;
293+
function setupTest(params: Partial<BetaToolRunnerParams> & { stream: true }): SetupTestResult<true>;
294+
function setupTest(params: Partial<BetaToolRunnerParams> = {}): SetupTestResult<boolean> {
296295
const { handleRequest, handleStreamEvents, fetch } = mockFetch();
297296
let messageIdCounter = 0;
298297
const handleAssistantMessage: SetupTestResult<false>['handleAssistantMessage'] = (
@@ -380,7 +379,7 @@ function setupTest(params: Partial<ToolRunnerParams> = {}): SetupTestResult<bool
380379

381380
const client = new OpenAI({ apiKey: 'test-key', fetch: fetch, maxRetries: 0 });
382381

383-
const runnerParams: ToolRunnerParams = {
382+
const runnerParams: BetaToolRunnerParams = {
384383
messages: params.messages || [{ role: 'user', content: 'What is the weather?' }],
385384
model: params.model || 'gpt-4o',
386385
max_tokens: params.max_tokens || 1000,

0 commit comments

Comments
 (0)