Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 10 additions & 0 deletions pnpm-lock.yaml

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

1 change: 1 addition & 0 deletions report-app/package.json
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@
"@shikijs/themes": "3.13.0",
"express": "^4.18.2",
"jszip": "^3.10.1",
"ngx-json-viewer": "^3.2.1",
"rxjs": "~7.8.0",
"shiki": "^3.6.0",
"tinyglobby": "^0.2.14",
Expand Down
26 changes: 26 additions & 0 deletions report-app/src/app/pages/report-viewer/report-viewer.html
Original file line number Diff line number Diff line change
Expand Up @@ -411,6 +411,32 @@ <h4>Debugging Tools</h4>
(click)="downloadDebuggingZip(result)">
Download ZIP for debugging
</button>
@if (result.toolLogs && result.toolLogs.length > 0) {
<expansion-panel>
<expansion-panel-header>Tool Logs</expansion-panel-header>
<ul class="tool-logs-list">
@for (log of result.toolLogs; track $index) {
<li>
<details class="details mcp-log-entry">
@let name = log.request.name;
<summary>
Log Entry #{{ $index + 1
}}{{ name ? ' - ' + name : '' }}
</summary>
<div class="mcp-log-content">
<h5>Request</h5>
<ngx-json-viewer [json]="log.request" [expanded]="false"></ngx-json-viewer>
<h5>Response</h5>
<ngx-json-viewer [json]="log.response" [expanded]="false"></ngx-json-viewer>
</div>
</details>
</li>
} @empty {
<li>No MCP logs were recorded for this run.</li>
}
</ul>
</expansion-panel>
}
</div>

@if (finalBuild.runtimeErrors) {
Expand Down
31 changes: 30 additions & 1 deletion report-app/src/app/pages/report-viewer/report-viewer.scss
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ expansion-panel {
padding: 0 1rem 1rem;
}

.app-details-section expansion-panel {
.app-details-section expansion-panel, .app-details-section button {
margin-bottom: 0.5rem;
}

Expand Down Expand Up @@ -233,3 +233,32 @@ expansion-panel {
padding: 0px 20px;
}

.mcp-log-entry {
border: 1px solid var(--border-color);
border-radius: var(--border-radius);
margin: 1rem 0;

& > summary {
cursor: pointer;
font-weight: 500;
padding: 1rem 1.5rem;

&:hover {
background-color: var(--button-active-bg-color);
}
}

& .mcp-log-content {
padding: 0 1.5rem 1.5rem;

h5 {
margin-top: 1.5rem;
margin-bottom: 0.5rem;
}
}
}

.tool-logs-list {
list-style: none;
padding: 0;
}
2 changes: 2 additions & 0 deletions report-app/src/app/pages/report-viewer/report-viewer.ts
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,7 @@ import {
signal,
viewChild,
} from '@angular/core';
import { NgxJsonViewerModule } from 'ngx-json-viewer';
import { BuildErrorType } from '../../../../../runner/builder/builder-types';
import {
AssessmentResult,
Expand Down Expand Up @@ -55,6 +56,7 @@ import { ProviderLabel } from '../../shared/provider-label';
ExpansionPanel,
ExpansionPanelHeader,
ProviderLabel,
NgxJsonViewerModule,
],
templateUrl: './report-viewer.html',
styleUrls: ['./report-viewer.scss'],
Expand Down
2 changes: 1 addition & 1 deletion runner/codegen/gemini-cli/gemini-cli-runner.ts
Original file line number Diff line number Diff line change
Expand Up @@ -89,7 +89,7 @@ export class GeminiCliRunner implements LlmRunner {
});
}

return { files, reasoning };
return { files, reasoning, toolLogs: [] };
}

generateText(): Promise<LlmGenerateTextResponse> {
Expand Down
51 changes: 50 additions & 1 deletion runner/codegen/genkit/genkit-runner.ts
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
import {
DynamicResourceAction,
GenerateResponse,
genkit,
ModelReference,
ToolAction,
Expand Down Expand Up @@ -27,6 +28,7 @@ import {
GenkitModelProvider,
PromptDataForCounting,
} from './model-provider.js';
import { ToolLogEntry } from '../../shared-interfaces.js';

const globalLogger = new GenkitLogger();
logger.init(globalLogger);
Expand All @@ -38,6 +40,7 @@ export class GenkitRunner implements LlmRunner {
readonly hasBuiltInRepairLoop = false;
private readonly genkitInstance = this.getGenkitInstance();
private mcpHost: GenkitMcpHost | null = null;
private toolLogs: ToolLogEntry[] = [];

async generateConstrained<T extends z.ZodTypeAny = z.ZodTypeAny>(
options: LlmConstrainedOutputGenerateRequestOptions<T>
Expand Down Expand Up @@ -75,9 +78,14 @@ export class GenkitRunner implements LlmRunner {
files: result.output.outputFiles || [],
usage: result.usage,
reasoning: result.reasoning,
toolLogs: this.flushToolLogs(),
};
}

flushToolLogs(): ToolLogEntry[] {
return this.toolLogs.splice(0);
}

async generateText(
options: LlmGenerateTextRequestOptions
): Promise<LlmGenerateTextResponse> {
Expand All @@ -87,6 +95,7 @@ export class GenkitRunner implements LlmRunner {
text: result.text,
usage: result.usage,
reasoning: result.reasoning,
toolLogs: this.flushToolLogs(),
};
}

Expand Down Expand Up @@ -120,7 +129,7 @@ export class GenkitRunner implements LlmRunner {
]);
}

return this.genkitInstance.generate({
const response = await this.genkitInstance.generate({
prompt: options.prompt,
model,
output: schema
Expand All @@ -145,6 +154,10 @@ export class GenkitRunner implements LlmRunner {
resources,
abortSignal: options.abortSignal,
});

this._logToolUsage(response);

return response;
};

return options.timeout
Expand All @@ -158,6 +171,42 @@ export class GenkitRunner implements LlmRunner {
);
}

private _logToolUsage(response: GenerateResponse<any>) {
const toolRequests = new Map<string, any>();
const toolResponses = new Map<string, any>();

if (response.request?.messages) {
for (const message of response.request.messages) {
if (!message.content) {
continue;
}
for (const contentPart of message.content) {
if (contentPart.toolRequest) {
toolRequests.set(
contentPart.toolRequest.ref || '0',
contentPart.toolRequest
);
} else if (contentPart.toolResponse) {
toolResponses.set(
contentPart.toolResponse.ref || '0',
contentPart.toolResponse
);
}
}
}
}

for (const [ref, toolRequest] of toolRequests.entries()) {
const toolResponse = toolResponses.get(ref);
if (toolResponse) {
this.toolLogs.push({
request: toolRequest,
response: toolResponse,
});
}
}
}

startMcpServerHost(hostName: string, servers: McpServerOptions[]): void {
if (this.mcpHost !== null) {
throw new Error('MCP host is already started');
Expand Down
20 changes: 11 additions & 9 deletions runner/codegen/llm-runner.ts
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
import { z } from 'zod';
import { LlmResponseFile, Usage } from '../shared-interfaces.js';
import { LlmResponseFile, ToolLogEntry, Usage } from '../shared-interfaces.js';
import { UserFacingError } from '../utils/errors.js';

export function assertValidModelName(value: string, availableModels: string[]) {
Expand Down Expand Up @@ -141,22 +141,24 @@ export interface LlmConstrainedOutputGenerateResponse<
reasoning: string;
}

/** File generation response from the LLM. */
export interface LlmGenerateFilesResponse {
files: LlmResponseFile[];
/** LLM response. */
interface BaseLlmGenerateResponse {
/** Token usage data, if available. */
usage?: Partial<Usage>;
/** Reasoning messages from the LLM. */
reasoning: string;
/** Tool requests and responses. */
toolLogs?: ToolLogEntry[];
}

/** File generation response from the LLM. */
export interface LlmGenerateFilesResponse extends BaseLlmGenerateResponse {
files: LlmResponseFile[];
}

/** Text response from the LLM. */
export interface LlmGenerateTextResponse {
export interface LlmGenerateTextResponse extends BaseLlmGenerateResponse {
text: string;
/** Token usage data, if available. */
usage?: Partial<Usage>;
/** Reasoning messages from the LLM. */
reasoning: string;
}

/** Schema for the LLM server options. */
Expand Down
17 changes: 8 additions & 9 deletions runner/orchestration/build.ts
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ import {
BuildWorkerMessage,
RepairType,
} from '../builder/builder-types.js';
import { LlmRunner } from '../codegen/llm-runner.js';
import { LlmGenerateFilesResponse, LlmRunner } from '../codegen/llm-runner.js';
import { Environment } from '../configuration/environment.js';
import {
AttemptDetails,
Expand Down Expand Up @@ -44,11 +44,7 @@ export async function attemptBuild(
rootPromptDef: RootPromptDefinition,
directory: string,
contextFiles: LlmContextFile[],
initialResponse: {
usage: Usage;
outputFiles: LlmResponseFile[];
reasoning: string;
},
initialResponse: LlmGenerateFilesResponse,
attemptDetails: AttemptDetails[],
skipScreenshots: boolean,
skipAxeTesting: boolean,
Expand All @@ -72,7 +68,7 @@ export async function attemptBuild(

// Clone the original files, because we're going to mutate them between repair
// attempts and we don't want the different runs to influence each other.
const finalOutputFiles = initialResponse.outputFiles.map((file) => ({
const finalOutputFiles = initialResponse.files.map((file) => ({
...file,
}));
let buildResult = await workerConcurrencyQueue.add(
Expand All @@ -86,8 +82,11 @@ export async function attemptBuild(
: DEFAULT_MAX_REPAIR_ATTEMPTS;

attemptDetails.push({
outputFiles: initialResponse.outputFiles,
usage: initialResponse.usage,
outputFiles: initialResponse.files,
usage: {
...{ inputTokens: 0, outputTokens: 0, totalTokens: 0 },
...initialResponse.usage,
},
reasoning: initialResponse.reasoning,
buildResult,
attempt: 0,
Expand Down
5 changes: 5 additions & 0 deletions runner/orchestration/codegen.ts
Original file line number Diff line number Diff line change
Expand Up @@ -3,6 +3,7 @@ import {
LlmResponse,
LlmResponseFile,
RootPromptDefinition,
ToolLogEntry,
Usage,
} from '../shared-interfaces.js';
import {
Expand Down Expand Up @@ -49,6 +50,7 @@ export async function generateCodeWithAI(
let usage: Usage;
let success: boolean;
let reasoning: string;
let toolLogs: ToolLogEntry[];

const contextMessageData = prepareContextFilesMessage(contextFiles);
const messages: PromptDataMessage[] | undefined = contextMessageData
Expand All @@ -72,6 +74,7 @@ export async function generateCodeWithAI(
totalTokens: response.usage?.totalTokens ?? 0,
};
reasoning = response.reasoning;
toolLogs = response.toolLogs ?? [];

progress.log(
promptDef,
Expand Down Expand Up @@ -100,6 +103,7 @@ export async function generateCodeWithAI(
usage = { inputTokens: 0, outputTokens: 0, totalTokens: 0 };
success = false;
reasoning = '';
toolLogs = [];
errors.push(error + '');
progress.log(
promptDef,
Expand All @@ -117,6 +121,7 @@ export async function generateCodeWithAI(
errors,
usage,
reasoning,
toolLogs,
} satisfies LlmResponse;
}

Expand Down
Loading
Loading