Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
89 changes: 50 additions & 39 deletions packages/sdk/server-ai/__tests__/LDAIConfigTrackerImpl.test.ts
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
import { LDContext } from '@launchdarkly/js-server-sdk-common';

import { name as aiSdkName, version as aiSdkVersion } from '../package.json';
import { LDFeedbackKind } from '../src/api/metrics';
import { LDAIConfigTrackerImpl } from '../src/LDAIConfigTrackerImpl';
import { LDClientMin } from '../src/LDClientMin';
Expand All @@ -18,6 +19,16 @@ const version = 1;
const modelName = 'test-model';
const providerName = 'test-provider';

const getExpectedTrackData = () => ({
configKey,
variationKey,
version,
modelName,
providerName,
aiSdkName,
aiSdkVersion,
});

beforeEach(() => {
jest.clearAllMocks();
});
Expand All @@ -37,7 +48,7 @@ it('tracks duration', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:duration:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1000,
);
});
Expand All @@ -60,7 +71,7 @@ it('tracks duration of async function', async () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:duration:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1000,
);
});
Expand All @@ -80,7 +91,7 @@ it('tracks time to first token', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:ttf',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1000,
);
});
Expand All @@ -100,7 +111,7 @@ it('tracks positive feedback', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:feedback:user:positive',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);
});
Expand All @@ -120,7 +131,7 @@ it('tracks negative feedback', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:feedback:user:negative',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);
});
Expand All @@ -140,7 +151,7 @@ it('tracks success', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:success',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);
});
Expand Down Expand Up @@ -172,14 +183,14 @@ it('tracks OpenAI usage', async () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:duration:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1000,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:success',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

Expand All @@ -193,21 +204,21 @@ it('tracks OpenAI usage', async () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
TOTAL_TOKENS,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:input',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
PROMPT_TOKENS,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:output',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
COMPLETION_TOKENS,
);
});
Expand All @@ -234,14 +245,14 @@ it('tracks error when OpenAI metrics function throws', async () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:duration:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1000,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:error',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

Expand Down Expand Up @@ -283,7 +294,7 @@ it('tracks Bedrock conversation with successful response', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:success',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

Expand All @@ -297,28 +308,28 @@ it('tracks Bedrock conversation with successful response', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:duration:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
500,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
TOTAL_TOKENS,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:input',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
PROMPT_TOKENS,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:output',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
COMPLETION_TOKENS,
);
});
Expand All @@ -345,7 +356,7 @@ it('tracks Bedrock conversation with error response', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:error',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

Expand Down Expand Up @@ -385,14 +396,14 @@ describe('Vercel AI SDK generateText', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:duration:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1000,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:success',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

Expand All @@ -406,21 +417,21 @@ describe('Vercel AI SDK generateText', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
TOTAL_TOKENS,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:input',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
PROMPT_TOKENS,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:output',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
COMPLETION_TOKENS,
);
});
Expand All @@ -447,14 +458,14 @@ describe('Vercel AI SDK generateText', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:duration:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1000,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:error',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

Expand Down Expand Up @@ -491,21 +502,21 @@ it('tracks tokens', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
TOTAL_TOKENS,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:input',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
PROMPT_TOKENS,
);

expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:output',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
COMPLETION_TOKENS,
);
});
Expand Down Expand Up @@ -537,7 +548,7 @@ it('only tracks non-zero token counts', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:input',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
50,
);

Expand Down Expand Up @@ -623,7 +634,7 @@ it('tracks duration when async function throws', async () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:duration:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1000,
);
});
Expand All @@ -643,7 +654,7 @@ it('tracks error', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:error',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);
});
Expand Down Expand Up @@ -679,27 +690,27 @@ describe('trackMetricsOf', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:success',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

// Should track token usage
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:total',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
100,
);
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:input',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
50,
);
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:tokens:output',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
50,
);
});
Expand Down Expand Up @@ -729,7 +740,7 @@ describe('trackMetricsOf', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:error',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);
});
Expand All @@ -755,7 +766,7 @@ describe('trackMetricsOf', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:error',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

Expand Down Expand Up @@ -789,7 +800,7 @@ describe('trackMetricsOf', () => {
expect(mockTrack).toHaveBeenCalledWith(
'$ld:ai:generation:success',
testContext,
{ configKey, variationKey, version, modelName, providerName },
getExpectedTrackData(),
1,
);

Expand Down
4 changes: 2 additions & 2 deletions packages/sdk/server-ai/package.json
Original file line number Diff line number Diff line change
Expand Up @@ -7,8 +7,8 @@
"type": "git",
"url": "https://github.com/launchdarkly/js-core.git"
},
"main": "dist/index.js",
"types": "dist/index.d.ts",
"main": "dist/src/index.js",
"types": "dist/src/index.d.ts",
"type": "commonjs",
"scripts": {
"build": "npx tsc",
Expand Down
5 changes: 5 additions & 0 deletions packages/sdk/server-ai/src/LDAIConfigTrackerImpl.ts
Original file line number Diff line number Diff line change
@@ -1,5 +1,6 @@
import { LDContext } from '@launchdarkly/js-server-sdk-common';

import { name as aiSdkName, version as aiSdkVersion } from '../package.json';
import { LDAIConfigTracker } from './api/config';
import { LDAIMetricSummary } from './api/config/LDAIConfigTracker';
import { EvalScore } from './api/judge/types';
Expand Down Expand Up @@ -32,13 +33,17 @@ export class LDAIConfigTrackerImpl implements LDAIConfigTracker {
version: number;
modelName: string;
providerName: string;
aiSdkName: string;
aiSdkVersion: string;
} {
return {
variationKey: this._variationKey,
configKey: this._configKey,
version: this._version,
modelName: this._modelName,
providerName: this._providerName,
aiSdkName,
aiSdkVersion,
};
}

Expand Down
Loading