Skip to content

Commit 5c355a5

Browse files
author
AWS
committed
Amazon SageMaker Service Update: This release adds APIs for new features for SageMaker endpoint to scale down to zero instances, native support for multi-adapter inference, and endpoint scaling improvements.
1 parent 152e409 commit 5c355a5

File tree

2 files changed

+71
-11
lines changed

2 files changed

+71
-11
lines changed
Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,6 @@
1+
{
2+
"type": "feature",
3+
"category": "Amazon SageMaker Service",
4+
"contributor": "",
5+
"description": "This release adds APIs for new features for SageMaker endpoint to scale down to zero instances, native support for multi-adapter inference, and endpoint scaling improvements."
6+
}

services/sagemaker/src/main/resources/codegen-resources/service-2.json

Lines changed: 65 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -7358,7 +7358,8 @@
73587358
"OnStartDeepHealthChecks":{
73597359
"shape":"OnStartDeepHealthChecks",
73607360
"documentation":"<p>A flag indicating whether deep health checks should be performed when the cluster instance group is created or updated.</p>"
7361-
}
7361+
},
7362+
"OverrideVpcConfig":{"shape":"VpcConfig"}
73627363
},
73637364
"documentation":"<p>Details of an instance group in a SageMaker HyperPod cluster.</p>"
73647365
},
@@ -7413,7 +7414,8 @@
74137414
"OnStartDeepHealthChecks":{
74147415
"shape":"OnStartDeepHealthChecks",
74157416
"documentation":"<p>A flag indicating whether deep health checks should be performed when the cluster instance group is created or updated.</p>"
7416-
}
7417+
},
7418+
"OverrideVpcConfig":{"shape":"VpcConfig"}
74177419
},
74187420
"documentation":"<p>The specifications of an instance group that you need to define.</p>"
74197421
},
@@ -7603,6 +7605,7 @@
76037605
"shape":"ClusterLifeCycleConfig",
76047606
"documentation":"<p>The LifeCycle configuration applied to the instance.</p>"
76057607
},
7608+
"OverrideVpcConfig":{"shape":"VpcConfig"},
76067609
"ThreadsPerCore":{
76077610
"shape":"ClusterThreadsPerCore",
76087611
"documentation":"<p>The number of threads per CPU core you specified under <code>CreateCluster</code>.</p>"
@@ -9688,9 +9691,7 @@
96889691
"required":[
96899692
"InferenceComponentName",
96909693
"EndpointName",
9691-
"VariantName",
9692-
"Specification",
9693-
"RuntimeConfig"
9694+
"Specification"
96949695
],
96959696
"members":{
96969697
"InferenceComponentName":{
@@ -21358,7 +21359,7 @@
2135821359
"ImageVersionArn":{
2135921360
"type":"string",
2136021361
"max":256,
21361-
"pattern":"^arn:aws(-[\\w]+)*:sagemaker:.+:[0-9]{12}:image-version/[a-z0-9]([-.]?[a-z0-9])*/[0-9]+$"
21362+
"pattern":"^(arn:aws(-[\\w]+)*:sagemaker:.+:[0-9]{12}:image-version/[a-z0-9]([-.]?[a-z0-9])*/[0-9]+|None)$"
2136221363
},
2136321364
"ImageVersionNumber":{
2136421365
"type":"integer",
@@ -21496,7 +21497,7 @@
2149621497
"documentation":"<p>The maximum MB of memory to allocate to run a model that you assign to an inference component.</p>"
2149721498
}
2149821499
},
21499-
"documentation":"<p>Defines the compute resources to allocate to run a model that you assign to an inference component. These resources include CPU cores, accelerators, and memory.</p>"
21500+
"documentation":"<p>Defines the compute resources to allocate to run a model, plus any adapter models, that you assign to an inference component. These resources include CPU cores, accelerators, and memory.</p>"
2150021501
},
2150121502
"InferenceComponentContainerSpecification":{
2150221503
"type":"structure",
@@ -21580,7 +21581,6 @@
2158021581
},
2158121582
"InferenceComponentSpecification":{
2158221583
"type":"structure",
21583-
"required":["ComputeResourceRequirements"],
2158421584
"members":{
2158521585
"ModelName":{
2158621586
"shape":"ModelName",
@@ -21596,7 +21596,11 @@
2159621596
},
2159721597
"ComputeResourceRequirements":{
2159821598
"shape":"InferenceComponentComputeResourceRequirements",
21599-
"documentation":"<p>The compute resources allocated to run the model assigned to the inference component.</p>"
21599+
"documentation":"<p>The compute resources allocated to run the model, plus any adapter models, that you assign to the inference component.</p> <p>Omit this parameter if your request is meant to create an adapter inference component. An adapter inference component is loaded by a base inference component, and it uses the compute resources of the base inference component.</p>"
21600+
},
21601+
"BaseInferenceComponentName":{
21602+
"shape":"InferenceComponentName",
21603+
"documentation":"<p>The name of an existing inference component that is to contain the inference component that you're creating with your request.</p> <p>Specify this parameter only if your request is meant to create an adapter inference component. An adapter inference component contains the path to an adapter model. The purpose of the adapter model is to tailor the inference output of a base foundation model, which is hosted by the base inference component. The adapter inference component uses the compute resources that you assigned to the base inference component.</p> <p>When you create an adapter inference component, use the <code>Container</code> parameter to specify the location of the adapter artifacts. In the parameter value, use the <code>ArtifactUrl</code> parameter of the <code>InferenceComponentContainerSpecification</code> data type.</p> <p>Before you can create an adapter inference component, you must have an existing inference component that contains the foundation model that you want to adapt.</p>"
2160021604
}
2160121605
},
2160221606
"documentation":"<p>Details about the resources to deploy with this inference component, including the model, container, and compute resources.</p>"
@@ -21618,7 +21622,11 @@
2161821622
},
2161921623
"ComputeResourceRequirements":{
2162021624
"shape":"InferenceComponentComputeResourceRequirements",
21621-
"documentation":"<p>The compute resources allocated to run the model assigned to the inference component.</p>"
21625+
"documentation":"<p>The compute resources allocated to run the model, plus any adapter models, that you assign to the inference component.</p>"
21626+
},
21627+
"BaseInferenceComponentName":{
21628+
"shape":"InferenceComponentName",
21629+
"documentation":"<p>The name of the base inference component that contains this inference component.</p>"
2162221630
}
2162321631
},
2162421632
"documentation":"<p>Details about the resources that are deployed with this inference component.</p>"
@@ -27099,7 +27107,7 @@
2709927107
},
2710027108
"ManagedInstanceScalingMinInstanceCount":{
2710127109
"type":"integer",
27102-
"min":1
27110+
"min":0
2710327111
},
2710427112
"ManagedInstanceScalingStatus":{
2710527113
"type":"string",
@@ -28955,6 +28963,20 @@
2895528963
"type":"integer",
2895628964
"min":0
2895728965
},
28966+
"ModelShardingConfig":{
28967+
"type":"structure",
28968+
"members":{
28969+
"Image":{
28970+
"shape":"OptimizationContainerImage",
28971+
"documentation":"<p>The URI of an LMI DLC in Amazon ECR. SageMaker uses this image to run the optimization.</p>"
28972+
},
28973+
"OverrideEnvironment":{
28974+
"shape":"OptimizationJobEnvironmentVariables",
28975+
"documentation":"<p>Environment variables that override the default ones in the model container.</p>"
28976+
}
28977+
},
28978+
"documentation":"<p>Settings for the model sharding technique that's applied by a model optimization job.</p>"
28979+
},
2895828980
"ModelSortKey":{
2895928981
"type":"string",
2896028982
"enum":[
@@ -30427,6 +30449,10 @@
3042730449
"ModelCompilationConfig":{
3042830450
"shape":"ModelCompilationConfig",
3042930451
"documentation":"<p>Settings for the model compilation technique that's applied by a model optimization job.</p>"
30452+
},
30453+
"ModelShardingConfig":{
30454+
"shape":"ModelShardingConfig",
30455+
"documentation":"<p>Settings for the model sharding technique that's applied by a model optimization job.</p>"
3043030456
}
3043130457
},
3043230458
"documentation":"<p>Settings for an optimization technique that you apply with a model optimization job.</p>",
@@ -32153,6 +32179,24 @@
3215332179
"ml.c6i.16xlarge",
3215432180
"ml.c6i.24xlarge",
3215532181
"ml.c6i.32xlarge",
32182+
"ml.m6i.large",
32183+
"ml.m6i.xlarge",
32184+
"ml.m6i.2xlarge",
32185+
"ml.m6i.4xlarge",
32186+
"ml.m6i.8xlarge",
32187+
"ml.m6i.12xlarge",
32188+
"ml.m6i.16xlarge",
32189+
"ml.m6i.24xlarge",
32190+
"ml.m6i.32xlarge",
32191+
"ml.r6i.large",
32192+
"ml.r6i.xlarge",
32193+
"ml.r6i.2xlarge",
32194+
"ml.r6i.4xlarge",
32195+
"ml.r6i.8xlarge",
32196+
"ml.r6i.12xlarge",
32197+
"ml.r6i.16xlarge",
32198+
"ml.r6i.24xlarge",
32199+
"ml.r6i.32xlarge",
3215632200
"ml.g5.xlarge",
3215732201
"ml.g5.2xlarge",
3215832202
"ml.g5.4xlarge",
@@ -32169,6 +32213,14 @@
3216932213
"ml.g6.16xlarge",
3217032214
"ml.g6.24xlarge",
3217132215
"ml.g6.48xlarge",
32216+
"ml.g6e.xlarge",
32217+
"ml.g6e.2xlarge",
32218+
"ml.g6e.4xlarge",
32219+
"ml.g6e.8xlarge",
32220+
"ml.g6e.12xlarge",
32221+
"ml.g6e.16xlarge",
32222+
"ml.g6e.24xlarge",
32223+
"ml.g6e.48xlarge",
3217232224
"ml.p4d.24xlarge",
3217332225
"ml.c7g.large",
3217432226
"ml.c7g.xlarge",
@@ -32230,11 +32282,13 @@
3223032282
"ml.trn1.2xlarge",
3223132283
"ml.trn1.32xlarge",
3223232284
"ml.trn1n.32xlarge",
32285+
"ml.trn2.48xlarge",
3223332286
"ml.inf2.xlarge",
3223432287
"ml.inf2.8xlarge",
3223532288
"ml.inf2.24xlarge",
3223632289
"ml.inf2.48xlarge",
3223732290
"ml.p5.48xlarge",
32291+
"ml.p5e.48xlarge",
3223832292
"ml.m7i.large",
3223932293
"ml.m7i.xlarge",
3224032294
"ml.m7i.2xlarge",

0 commit comments

Comments
 (0)