This document is relevant for: Inf1, Inf2, Trn1, Trn2, Trn3

Neuron Compiler Error Codes#

This page lists the error codes you can encounter while developing with the Neuron Compiler. For more details on any individual error, click the link for that error code in the table below.

Error Code

Error Message

Recommendation

NCC_EARG001

Unsupported Logical Neuron Core (LNC) configuration.

You attempted to use a Logical Neuron Core configuration that is not supported by the target Neuron architecture.

NCC_EBVF030

The number of instructions generated exceeds the limit.

Consider applying model parallelism as partitioning the model will help break large computational graphs into smaller subgraphs.

NCC_EHCA005

The compiler encountered a custom call instruction with a target name that is not recognized.

Use a supported custom call target from the list of recognized targets.

NCC_EOOM001

The combined memory needed for the model’s activation tensors exceeds the high-bandwidth memory limit.

You may need to reduce batch/tensor size or utilize pipeline/tensor parallelism via neuronx-distributed.

NCC_EOOM002

The combined memory needed for the model’s activation tensors exceeds the high-bandwidth memory limit.

You may need to reduce batch/tensor size or utilize pipeline/tensor parallelism via neuronx-distributed.

NCC_ESFH002

The compiler encountered a unsigned 64-bit integer constant with a value that cannot be safely converted to 32-bit representation.

Try to use uint32 for constants when possible and restructure code to avoid large constants.

NCC_ESPP004

The compiler encountered a data type that is not supported for code generation.

Use a supported data type as listed in the Neuron documentation.

NCC_ESPP047

Unsupported 8-bit floating-point data type.

The compiler found usage of an unsupported 8-bit floating-point data type. Convert to a supported type like torch.float16.

NCC_EUOC002

An unsupported operator was used.

Try using alternative operators from the full list of supported operators via neuronx-cc list-operators –framework XLA to workaround the limitation.

NCC_EVRF001

An unsupported operator was used.

Try using alternative operators from the full list of supported operators to workaround the limitation.

NCC_EVRF004

Complex data types are not supported on the Neuron device.

You cannot use complex data types (such as complex64, complex128, and others) on the Neuron device directly.

NCC_EVRF005

Unsupported F8E4M3FNUZ, F8E4M3B11FNUZ, or F8E5M2FNUZ data type.

The compiler found usage of unsupported 8-bit floating-point data types. Convert to a supported type like torch.float16.

NCC_EVRF006

The compiler encountered a RNGBitGenerator operation using a random number generation algorithm other than RNG_DEFAULT.

Ensure that you are using standard JAX/PyTorch random APIs and not explicity specifying an RNG algorithm.

NCC_EVRF007

The number of instructions generated exceeds the limit.

Consider applying model parallelism as partitioning the model will help break large computational graphs into smaller subgraphs.

NCC_EVRF009

The combined memory needed for the model’s activation tensors exceeds the high-bandwidth memory limit.

You may need to reduce batch/tensor size or utilize pipeline/tensor parallelism via neuronx-distributed.

NCC_EVRF010

The compiler encountered simultaneous use of input and kernel dilation, which is not supported.

If possible, use only input or kernel dilation, not both simultaneously.

NCC_EVRF011

The compiler encountered strided convolution combined with dilated input, which is not supported.

If possible, remove stride or input dilation, or apply upsampling and downsampling separately.

NCC_EVRF013

TopK does not support integer input tensors (int32, int64).

The TopK operation cannot be performed on integer data types.

NCC_EVRF015

The compiler encountered a custom call instruction with a target name that is not recognized.

Use a supported custom call target from the list of recognized targets.

NCC_EVRF016

The scatter-reduce operation cannot perform reduction logic if the data being scattered or the destination tensor is using an integer or boolean data type.

Cast your input and source tensors to a floating-point data type (e.g., torch.float32 or torch.bfloat16).

NCC_EVRF017

Reduce-window operation with base dilation greater than 1 is not supported.

Change base dilation to be all 1s or consider manual dilation if necessary.

NCC_EVRF018

Reduce-window operation with window dilation greater than 1 is not supported.

Remove window_dilation or change values to be all 1s, or consider manual dilation if necessary.

NCC_EVRF019

The compiler encountered a reduce-window operation with more or less than 2 operands.

If possible, split multi-operand reduce_window with multiple single-operand reduce_window operations.

NCC_EVRF022

Shift-right-arithmetic operation on non 32-bit inputs is not supported. Cast the first argument’s data type to be S32, U32, or F32.

You need to use 32-bit data types for shift operations. Cast inputs to int32, uint32, or float32.

NCC_EVRF024

The output tensor size limit of 4GB was exceeded.

Reduce batch/tensor size or utilize tensor parallelism via neuronx-distributed.

NCC_EVRF031

The compiler encountered a scatter out-of-bounds error.

Ensure that the iota size matches the operand dimension size.

NCC_EXSP001

The combined memory needed for the model’s activation tensors exceeds the high-bandwidth memory limit.

You may need to reduce batch/tensor size or utilize pipeline/tensor parallelism via neuronx-distributed.

NCC_EXTP004

The number of instructions generated exceeds the limit.

Consider applying model parallelism as partitioning the model will help break large computational graphs into smaller subgraphs.

This document is relevant for: Inf1, Inf2, Trn1, Trn2, Trn3