This document is relevant for: Inf1
, Inf2
, Trn1
, Trn2
NeuronPerf Overview#
NeuronPerf is a lightweight Python library that can help you easily benchmark your models with Neuron hardware.
NeuronPerf supports Neuron releases for PyTorch, Tensorflow, and MXNet. It is used internally by the Neuron team to generate performance benchmarking numbers.
When interacting with NeuronPerf, you will typically import the base package along with one of the submodule wrappers, for example:
import neuronperf
import neuronperf.torch
You may then benchmark and/or compile one or more models with NeuronPerf. For example,
reports = neuronperf.torch.benchmark(model, inputs, ...)
The compile
and benchmark
methods must be accessed through one of the supported framework submodules.
Benchmarking#
All NeuronPerf benchmark
calls require a minimum of two arguments:
A filename
Inputs
The filename may refer to:
A Neuron-compiled model (e.g.
my_model.pt
)A Model Index.
A Model Index is useful for benchmarking more than one model in a single session.
Compiling#
NeuronPerf also provides a standard interface to all Neuron frameworks through the compile
API.
model_index = neuronperf.torch.compile(model, inputs, ...)
This is completely optional. You may use the standard compilation guides for supported frameworks.
Next Steps#
Take a look at the simple NeuronPerf Examples, NeuronPerf Benchmark Guide, NeuronPerf Compile Guide, and NeuronPerf API.
This document is relevant for: Inf1
, Inf2
, Trn1
, Trn2