Skip to content

Commit 1a787f4

Browse files
committed
fix: remove rating LLM option
Removes the option to customize the rating LLM runner. It didn't really make sense anymore, given that we only allow the Genkit one further downstream.
1 parent 932301a commit 1a787f4

File tree

2 files changed

+5
-16
lines changed

2 files changed

+5
-16
lines changed

runner/eval-cli.ts

Lines changed: 1 addition & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,5 @@
11
import { Arguments, Argv, CommandModule } from 'yargs';
22
import chalk from 'chalk';
3-
import { assertValidModelName, LlmRunner } from './codegen/llm-runner.js';
43
import {
54
BUILT_IN_ENVIRONMENTS,
65
DEFAULT_AUTORATER_MODEL_NAME,
@@ -11,8 +10,7 @@ import {
1110
logReportToConsole,
1211
writeReportToDisk,
1312
} from './reporting/report-logging.js';
14-
import { getRunnerByName, RunnerName } from './codegen/runner-creation.js';
15-
import { GenkitRunner } from './codegen/genkit/genkit-runner.js';
13+
import { RunnerName } from './codegen/runner-creation.js';
1614
import { UserFacingError } from './utils/errors.js';
1715

1816
export const EvalModule = {
@@ -166,9 +164,6 @@ function builder(argv: Argv): Argv<Options> {
166164
}
167165

168166
async function handler(cliArgs: Arguments<Options>): Promise<void> {
169-
let llm: LlmRunner | null = null;
170-
let ratingLlm: GenkitRunner | null = null;
171-
172167
if (!cliArgs.environment) {
173168
console.error(
174169
chalk.red(
@@ -184,9 +179,7 @@ async function handler(cliArgs: Arguments<Options>): Promise<void> {
184179
}
185180

186181
try {
187-
ratingLlm = await getRunnerByName('genkit');
188182
const runInfo = await generateCodeAndAssess({
189-
ratingLlm,
190183
runner: cliArgs.runner,
191184
model: cliArgs.model,
192185
environmentConfigPath:
@@ -222,7 +215,5 @@ async function handler(cliArgs: Arguments<Options>): Promise<void> {
222215
console.error(chalk.red((error as Error).stack));
223216
}
224217
}
225-
} finally {
226-
await ratingLlm?.dispose();
227218
}
228219
}

runner/orchestration/generate.ts

Lines changed: 4 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,6 @@ import {
99
assertValidModelName,
1010
LlmGenerateFilesContext,
1111
LlmGenerateFilesResponse,
12-
LlmRunner,
1312
} from '../codegen/llm-runner.js';
1413
import {
1514
DEFAULT_AUTORATER_MODEL_NAME,
@@ -54,9 +53,8 @@ import { UserFacingError } from '../utils/errors.js';
5453
import { getRunGroupId } from './grouping.js';
5554
import { executeCommand } from '../utils/exec.js';
5655
import { EvalID, Gateway } from './gateway.js';
57-
import { LocalGateway } from './gateways/local_gateway.js';
5856
import { LocalEnvironment } from '../configuration/environment-local.js';
59-
import { RunnerName } from '../codegen/runner-creation.js';
57+
import { getRunnerByName, RunnerName } from '../codegen/runner-creation.js';
6058

6159
/**
6260
* Orchestrates the entire assessment process for each prompt defined in the `prompts` array.
@@ -71,7 +69,6 @@ import { RunnerName } from '../codegen/runner-creation.js';
7169
* each containing the prompt, generated code, and final validation status.
7270
*/
7371
export async function generateCodeAndAssess(options: {
74-
ratingLlm: GenkitRunner;
7572
model: string;
7673
runner: RunnerName;
7774
environmentConfigPath: string;
@@ -96,6 +93,7 @@ export async function generateCodeAndAssess(options: {
9693
options.environmentConfigPath,
9794
options.runner
9895
);
96+
const ratingLlm = await getRunnerByName('genkit');
9997

10098
// TODO(devversion): Consider validating model names also for remote environments.
10199
if (env instanceof LocalEnvironment) {
@@ -179,7 +177,7 @@ export async function generateCodeAndAssess(options: {
179177
evalID,
180178
env,
181179
env.gateway,
182-
options.ratingLlm,
180+
ratingLlm,
183181
options.model,
184182
rootPromptDef,
185183
options.localMode,
@@ -254,7 +252,7 @@ export async function generateCodeAndAssess(options: {
254252
const timestamp = new Date();
255253
const details = {
256254
summary: await prepareSummary(
257-
options.ratingLlm,
255+
ratingLlm,
258256
new AbortController().signal, // Note: AI summarization is currently not abortable.
259257
options.model,
260258
env,

0 commit comments

Comments
 (0)