This document is relevant for: Inf1, Inf2, Trn1, Trn2, Trn3
Neuron Compiler Error Codes#
This page lists the error codes you can encounter while developing with the Neuron Compiler. For more details on any individual error, click the link for that error code in the table below.
Error Code |
Error Message |
Recommendation |
|---|---|---|
Unsupported Logical Neuron Core (LNC) configuration. |
You attempted to use a Logical Neuron Core configuration that is not supported by the target Neuron architecture. |
|
The number of instructions generated exceeds the limit. |
Consider applying model parallelism as partitioning the model will help break large computational graphs into smaller subgraphs. |
|
The compiler encountered a custom call instruction with a target name that is not recognized. |
Use a supported custom call target from the list of recognized targets. |
|
The combined memory needed for the model’s activation tensors exceeds the high-bandwidth memory limit. |
You may need to reduce batch/tensor size or utilize pipeline/tensor parallelism via neuronx-distributed. |
|
The combined memory needed for the model’s activation tensors exceeds the high-bandwidth memory limit. |
You may need to reduce batch/tensor size or utilize pipeline/tensor parallelism via neuronx-distributed. |
|
The compiler encountered a unsigned 64-bit integer constant with a value that cannot be safely converted to 32-bit representation. |
Try to use uint32 for constants when possible and restructure code to avoid large constants. |
|
The compiler encountered a data type that is not supported for code generation. |
Use a supported data type as listed in the Neuron documentation. |
|
Unsupported 8-bit floating-point data type. |
The compiler found usage of an unsupported 8-bit floating-point data type. Convert to a supported type like torch.float16. |
|
An unsupported operator was used. |
Try using alternative operators from the full list of supported operators via neuronx-cc list-operators –framework XLA to workaround the limitation. |
|
An unsupported operator was used. |
Try using alternative operators from the full list of supported operators to workaround the limitation. |
|
Complex data types are not supported on the Neuron device. |
You cannot use complex data types (such as |
|
Unsupported F8E4M3FNUZ, F8E4M3B11FNUZ, or F8E5M2FNUZ data type. |
The compiler found usage of unsupported 8-bit floating-point data types. Convert to a supported type like torch.float16. |
|
The compiler encountered a RNGBitGenerator operation using a random number generation algorithm other than RNG_DEFAULT. |
Ensure that you are using standard JAX/PyTorch random APIs and not explicity specifying an RNG algorithm. |
|
The number of instructions generated exceeds the limit. |
Consider applying model parallelism as partitioning the model will help break large computational graphs into smaller subgraphs. |
|
The combined memory needed for the model’s activation tensors exceeds the high-bandwidth memory limit. |
You may need to reduce batch/tensor size or utilize pipeline/tensor parallelism via neuronx-distributed. |
|
The compiler encountered simultaneous use of input and kernel dilation, which is not supported. |
If possible, use only input or kernel dilation, not both simultaneously. |
|
The compiler encountered strided convolution combined with dilated input, which is not supported. |
If possible, remove stride or input dilation, or apply upsampling and downsampling separately. |
|
TopK does not support integer input tensors (int32, int64). |
The TopK operation cannot be performed on integer data types. |
|
The compiler encountered a custom call instruction with a target name that is not recognized. |
Use a supported custom call target from the list of recognized targets. |
|
The scatter-reduce operation cannot perform reduction logic if the data being scattered or the destination tensor is using an integer or boolean data type. |
Cast your input and source tensors to a floating-point data type (e.g., torch.float32 or torch.bfloat16). |
|
Reduce-window operation with base dilation greater than 1 is not supported. |
Change base dilation to be all 1s or consider manual dilation if necessary. |
|
Reduce-window operation with window dilation greater than 1 is not supported. |
Remove window_dilation or change values to be all 1s, or consider manual dilation if necessary. |
|
The compiler encountered a reduce-window operation with more or less than 2 operands. |
If possible, split multi-operand reduce_window with multiple single-operand reduce_window operations. |
|
Shift-right-arithmetic operation on non 32-bit inputs is not supported. Cast the first argument’s data type to be S32, U32, or F32. |
You need to use 32-bit data types for shift operations. Cast inputs to int32, uint32, or float32. |
|
The output tensor size limit of 4GB was exceeded. |
Reduce batch/tensor size or utilize tensor parallelism via neuronx-distributed. |
|
The compiler encountered a scatter out-of-bounds error. |
Ensure that the iota size matches the operand dimension size. |
|
The combined memory needed for the model’s activation tensors exceeds the high-bandwidth memory limit. |
You may need to reduce batch/tensor size or utilize pipeline/tensor parallelism via neuronx-distributed. |
|
The number of instructions generated exceeds the limit. |
Consider applying model parallelism as partitioning the model will help break large computational graphs into smaller subgraphs. |
This document is relevant for: Inf1, Inf2, Trn1, Trn2, Trn3