GitHub / aws-samples / optimize-foundation-models-deployment-on-amazon-sagemaker
In this workshop, we demonstrate how to choose the right container and right instance types, optimize container parameters, and set up the right autoscaling policies and how to use APIs to get recommendations with Amazon SageMaker
Stars: 3
Forks: 0
Open issues: 1
License: mit-0
Language: Jupyter Notebook
Size: 1.25 MB
Dependencies parsed at: Pending
Created at: over 1 year ago
Updated at: 8 months ago
Pushed at: about 1 month ago
Last synced at: 5 days ago
Topics: llm-inference, sagemaker, sagemaker-deployment