Product

Team Member

Amazon Inferentia

IT Infrastructure
Other IT Infrastructure

The custom-designed chip for machine learning inference at scale.

☆☆☆☆☆ 0.0 Based on 0 Reviews

Amazon Inferentia

Learn More

About the Amazon Inferentia

Amazon Inferentia is a custom-built, high-performance machine learning inference chip designed by Amazon Web Services (AWS) to provide the lowest cost per inference in the cloud. It powers Amazon EC2 Inf1 and Inf2 instances, which are optimized for running machine learning models for applications like natural language processing, computer vision, and recommendation engines. By using the AWS Neuron SDK, developers can easily compile and run their machine learning models from popular frameworks like TensorFlow and PyTorch with minimal code changes, allowing for high throughput and low latency.

Supported Platforms

Supported Languages

Pricing Of Amazon Inferentia

Features Of Amazon Inferentia

Amazon Inferentia Media