Scale distributed machine learning (ML) training to thousands of accelerated instances and seamlessly deploy models for inference in production. Develop on accelerators—including AWS Trainium, AWS Inferentia, and NVIDIA GPUs—with the newest drivers, frameworks, libraries, and tools.