Replies: 13 comments
-
| 
         Hello @UniSabina, Apologies on the slow experience. I'll reach out to the team that handles the hosting platform and let them comment.  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         I'm also interested in this  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         Any update on this?  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         Hii @ChoiByungWook,  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         Is there any update on this? I'm experiencing a similar issue.  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         Any update about this issue? I am still having the same problem, the model deployment take too long.  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         It is too slow for training and deploying model via SageMaker Studio. I just tested with Iris dataset.  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         really slow  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         Still no update? I can build and push a custom PyTorch training image, and then train a model in a shorter time than deploying.  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         Takes over 90 minutes for me.  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         How much should I wait before killing the process?!  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         so slow still  | 
  
Beta Was this translation helpful? Give feedback.
-
| 
         terribly slow... tring to deploy an ml.m4.xlarge, since the max health_check_timeout is topped at 3600 and my deployment currently takes over an hour, I am not able to deploy llama3 70B. What to do?  | 
  
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
System Information
Describe the problem
I'm quite new to the SageMaker algorithms and estimators so please bear with me.
I'm running a script very similar to this example script for DeepAR
https://github.com/awslabs/amazon-sagemaker-examples/blob/master/introduction_to_amazon_algorithms/deepar_electricity/DeepAR-Electricity.ipynb
And want to start more than hundred of such training + prediction jobs.
The cell
takes up 70% (~8.5min) of the time of the overall training and predicting job (~12min). Is there a possibility to reduce that time? What is the reason for this deploy job taking so long?
Thanks!
Beta Was this translation helpful? Give feedback.
All reactions