Amazon SageMaker HyperPod now supports NVIDIA Multi-Instance GPU (MIG) for generative AI tasks

Amazon SageMaker HyperPod Update
Amazon SageMaker HyperPod now supports NVIDIA Multi-Instance GPU (MIG) technology, allowing administrators to partition a single GPU into multiple isolated GPUs. This enables running diverse, small generative AI tasks simultaneously while maintaining performance and task isolation.
Administrators can configure GPU partitions via the SageMaker HyperPod console or a custom setup. They can also allocate compute quota for fair distribution across teams. Real-time performance metrics and resource utilization monitoring are available to optimize resource allocation.
Data scientists can accelerate time-to-market by scheduling lightweight inference tasks and running interactive notebooks in parallel on GPU partitions.
Available Regions
- US West (Oregon)
- US East (N.Virginia)
- US East (Ohio)
- US West (N. California)
- Canada (Central)
- South America (Sao Paulo)
- Europe (Stockholm)
- Europe (Spain)
- Europe (Ireland)
- Europe (Frankfurt)
- Europe (London)
- Asia Pacific (Mumbai)
- Asia Pacific (Jakarta)
- Asia Pacific (Melbourne)
- Asia Pacific (Tokyo)
- Asia Pacific (Sydney)
- Asia Pacific (Seoul)
- Asia Pacific (Singapore)
What to do
- Visit the SageMaker HyperPod webpage for more information.
- Read the SageMaker HyperPod documentation to get started.
Source: AWS release notes
If you need further guidance on AWS, our experts are available at AWS@westloop.io. You may also reach us by submitting the Contact Us form.



