Skip to content

Commit b436013

Browse files
author
awstools
committed
feat(client-sagemaker): This release adds a new Neuron driver option in InferenceAmiVersion parameter for ProductionVariant. Additionally, it adds support for fetching model lifecycle status in the ListModelPackages API. Users can now use this API to view the lifecycle stage of models that have been shared with them.
1 parent 059c057 commit b436013

File tree

8 files changed

+76792
-76752
lines changed

8 files changed

+76792
-76752
lines changed

clients/client-sagemaker/src/commands/CreateEndpointConfigCommand.ts

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -96,7 +96,7 @@ export interface CreateEndpointConfigCommandOutput extends CreateEndpointConfigO
9696
* RoutingConfig: { // ProductionVariantRoutingConfig
9797
* RoutingStrategy: "LEAST_OUTSTANDING_REQUESTS" || "RANDOM", // required
9898
* },
99-
* InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1",
99+
* InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1" || "al2-ami-sagemaker-inference-neuron-2",
100100
* },
101101
* ],
102102
* DataCaptureConfig: { // DataCaptureConfig
@@ -209,7 +209,7 @@ export interface CreateEndpointConfigCommandOutput extends CreateEndpointConfigO
209209
* RoutingConfig: {
210210
* RoutingStrategy: "LEAST_OUTSTANDING_REQUESTS" || "RANDOM", // required
211211
* },
212-
* InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1",
212+
* InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1" || "al2-ami-sagemaker-inference-neuron-2",
213213
* },
214214
* ],
215215
* ExecutionRoleArn: "STRING_VALUE",

clients/client-sagemaker/src/commands/DescribeEndpointConfigCommand.ts

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ export interface DescribeEndpointConfigCommandOutput extends DescribeEndpointCon
7373
* // RoutingConfig: { // ProductionVariantRoutingConfig
7474
* // RoutingStrategy: "LEAST_OUTSTANDING_REQUESTS" || "RANDOM", // required
7575
* // },
76-
* // InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1",
76+
* // InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1" || "al2-ami-sagemaker-inference-neuron-2",
7777
* // },
7878
* // ],
7979
* // DataCaptureConfig: { // DataCaptureConfig
@@ -181,7 +181,7 @@ export interface DescribeEndpointConfigCommandOutput extends DescribeEndpointCon
181181
* // RoutingConfig: {
182182
* // RoutingStrategy: "LEAST_OUTSTANDING_REQUESTS" || "RANDOM", // required
183183
* // },
184-
* // InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1",
184+
* // InferenceAmiVersion: "al2-ami-sagemaker-inference-gpu-2" || "al2-ami-sagemaker-inference-gpu-2-1" || "al2-ami-sagemaker-inference-gpu-3-1" || "al2-ami-sagemaker-inference-neuron-2",
185185
* // },
186186
* // ],
187187
* // ExecutionRoleArn: "STRING_VALUE",

clients/client-sagemaker/src/commands/ListModelPackagesCommand.ts

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -60,6 +60,11 @@ export interface ListModelPackagesCommandOutput extends ListModelPackagesOutput,
6060
* // CreationTime: new Date("TIMESTAMP"), // required
6161
* // ModelPackageStatus: "Pending" || "InProgress" || "Completed" || "Failed" || "Deleting", // required
6262
* // ModelApprovalStatus: "Approved" || "Rejected" || "PendingManualApproval",
63+
* // ModelLifeCycle: { // ModelLifeCycle
64+
* // Stage: "STRING_VALUE", // required
65+
* // StageStatus: "STRING_VALUE", // required
66+
* // StageDescription: "STRING_VALUE",
67+
* // },
6368
* // },
6469
* // ],
6570
* // NextToken: "STRING_VALUE",

clients/client-sagemaker/src/models/models_0.ts

Lines changed: 22 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2292,7 +2292,28 @@ export interface StoppingCondition {
22922292

22932293
/**
22942294
* <p>The maximum length of time, in seconds, that a training or compilation job can be
2295-
* pending before it is stopped.</p>
2295+
* pending before it is stopped.</p>
2296+
* <note>
2297+
* <p>When working with training jobs that use capacity from <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/reserve-capacity-with-training-plans.html">training
2298+
* plans</a>, not all <code>Pending</code> job states count against the
2299+
* <code>MaxPendingTimeInSeconds</code> limit. The following scenarios do not
2300+
* increment the <code>MaxPendingTimeInSeconds</code> counter:</p>
2301+
* <ul>
2302+
* <li>
2303+
* <p>The plan is in a <code>Scheduled</code> state: Jobs queued (in
2304+
* <code>Pending</code> status) before a plan's start date (waiting for
2305+
* scheduled start time)</p>
2306+
* </li>
2307+
* <li>
2308+
* <p>Between capacity reservations: Jobs temporarily back to
2309+
* <code>Pending</code> status between two capacity reservation
2310+
* periods</p>
2311+
* </li>
2312+
* </ul>
2313+
* <p>
2314+
* <code>MaxPendingTimeInSeconds</code> only increments when jobs are actively
2315+
* waiting for capacity in an <code>Active</code> plan.</p>
2316+
* </note>
22962317
* @public
22972318
*/
22982319
MaxPendingTimeInSeconds?: number | undefined;

clients/client-sagemaker/src/models/models_1.ts

Lines changed: 12 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4455,6 +4455,7 @@ export const ProductionVariantInferenceAmiVersion = {
44554455
AL2_GPU_2: "al2-ami-sagemaker-inference-gpu-2",
44564456
AL2_GPU_2_1: "al2-ami-sagemaker-inference-gpu-2-1",
44574457
AL2_GPU_3_1: "al2-ami-sagemaker-inference-gpu-3-1",
4458+
AL2_NEURON_2: "al2-ami-sagemaker-inference-neuron-2",
44584459
} as const;
44594460

44604461
/**
@@ -4745,6 +4746,17 @@ export interface ProductionVariant {
47454746
* </li>
47464747
* </ul>
47474748
* </dd>
4749+
* <dt>al2-ami-sagemaker-inference-neuron-2</dt>
4750+
* <dd>
4751+
* <ul>
4752+
* <li>
4753+
* <p>Accelerator: Inferentia2 and Trainium</p>
4754+
* </li>
4755+
* <li>
4756+
* <p>Neuron driver version: 2.19</p>
4757+
* </li>
4758+
* </ul>
4759+
* </dd>
47484760
* </dl>
47494761
* @public
47504762
*/

clients/client-sagemaker/src/models/models_4.ts

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -3479,6 +3479,12 @@ export interface ModelPackageSummary {
34793479
* @public
34803480
*/
34813481
ModelApprovalStatus?: ModelApprovalStatus | undefined;
3482+
3483+
/**
3484+
* <p> A structure describing the current state of the model in its life cycle. </p>
3485+
* @public
3486+
*/
3487+
ModelLifeCycle?: ModelLifeCycle | undefined;
34823488
}
34833489

34843490
/**

clients/client-sagemaker/src/protocols/Aws_json1_1.ts

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -27207,6 +27207,7 @@ const de_ModelPackageSummary = (output: any, context: __SerdeContext): ModelPack
2720727207
return take(output, {
2720827208
CreationTime: (_: any) => __expectNonNull(__parseEpochTimestamp(__expectNumber(_))),
2720927209
ModelApprovalStatus: __expectString,
27210+
ModelLifeCycle: (_: any) => de_ModelLifeCycle(_, context),
2721027211
ModelPackageArn: __expectString,
2721127212
ModelPackageDescription: __expectString,
2721227213
ModelPackageGroupName: __expectString,

0 commit comments

Comments
 (0)