Skip to content

Commit 29a6262

Browse files
feat(inference): remove DeploymentQuantization.enabled field (#1938)
Co-authored-by: Jonathan R. <[email protected]>
1 parent b981ef7 commit 29a6262

File tree

2 files changed

+0
-4
lines changed

2 files changed

+0
-4
lines changed

packages/clients/src/api/inference/v1/marshalling.gen.ts

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -132,7 +132,6 @@ const unmarshalDeploymentQuantization = (
132132

133133
return {
134134
bits: data.bits,
135-
enabled: data.enabled,
136135
} as DeploymentQuantization
137136
}
138137

@@ -279,7 +278,6 @@ const marshalDeploymentQuantization = (
279278
defaults: DefaultValues,
280279
): Record<string, unknown> => ({
281280
bits: request.bits,
282-
enabled: request.enabled,
283281
})
284282

285283
const marshalEndpointSpec = (

packages/clients/src/api/inference/v1/types.gen.ts

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -63,8 +63,6 @@ export interface ModelSupportedNode {
6363
}
6464

6565
export interface DeploymentQuantization {
66-
/** Whether to enable quantization for this deployment. */
67-
enabled: boolean
6866
/**
6967
* The number of bits each model parameter should be quantized to. The
7068
* quantization method is chosen based on this value.

0 commit comments

Comments
 (0)