Main Content

Supported Networks and Layers

Supported Pretrained Networks

GPU Coder™ supports code generation for series and directed acyclic graph (DAG) convolutional neural networks (CNNs or ConvNets). You can generate code for any trained convolutional neural network whose layers are supported for code generation. SeeSupported Layers. You can train a convolutional neural network on either a CPU, a GPU, or multiple GPUs by using the Deep Learning Toolbox™ or use one of the pretrained networks listed in the table and generate CUDA®code.

Network Name Description cuDNN TensorRT ARM®Compute Library for Mali GPU

AlexNet

AlexNet convolutional neural network. For the pretrained AlexNet model, seealexnet(Deep Learning Toolbox).

The syntaxalexnet('Weights','none')is not supported for code generation.

Yes

Yes

Yes

Caffe Network

Convolutional neural network models from Caffe. For importing a pretrained network from Caffe, seeimportCaffeNetwork(Deep Learning Toolbox).

Yes

Yes

Yes

Darknet-19

Darknet-19 convolutional neural network. For more information, seedarknet19(Deep Learning Toolbox).

The syntaxdarknet19('Weights','none')is not supported for code generation.

Yes

Yes

Yes

Darknet-53

Darknet-53 convolutional neural network. for more information, seedarknet53(Deep Learning Toolbox).

The syntaxdarknet53('Weights','none')is not supported for code generation.

Yes

Yes

Yes

DeepLab v3+

DeepLab v3+ convolutional neural network. For more information, seedeeplabv3plusLayers(Computer Vision Toolbox).

Yes

Yes

No

DenseNet-201

DenseNet-201 convolutional neural network. For the pretrained DenseNet-201 model, seedensenet201(Deep Learning Toolbox).

The syntaxdensenet201('Weights','none')is not supported for code generation.

Yes

Yes

Yes

EfficientNet-b0

EfficientNet-b0 convolutional neural network. For the pretrained EfficientNet-b0 model, seeefficientnetb0(Deep Learning Toolbox).

The syntaxefficientnetb0('Weights','none')is not supported for code generation.

Yes Yes Yes

GoogLeNet

GoogLeNet convolutional neural network. For the pretrained GoogLeNet model, seegooglenet(Deep Learning Toolbox).

The syntaxgooglenet('Weights','none')is not supported for code generation.

Yes

Yes

Yes

Inception-ResNet-v2

Inception-ResNet-v2 convolutional neural network. For the pretrained Inception-ResNet-v2 model, see在ceptionresnetv2(Deep Learning Toolbox).

Yes

Yes

No

Inception-v3

Inception-v3 convolutional neural network. For the pretrained Inception-v3 model, see在ceptionv3(Deep Learning Toolbox).

The syntax在ceptionv3('Weights','none')is not supported for code generation.

Yes

Yes

Yes

Mobilenet-v2

MobileNet-v2 convolutional neural network. For the pretrained MobileNet-v2 model, seemobilenetv2(Deep Learning Toolbox).

The syntaxmobilenetv2('Weights','none')is not supported for code generation.

Yes

Yes

Yes

NASNet-Large

NASNet-Large convolutional neural network. For the pretrained NASNet-Large model, seenasnetlarge(Deep Learning Toolbox).

Yes

Yes

No

NASNet-Mobile

NASNet-Mobile convolutional neural network. For the pretrained NASNet-Mobile model, seenasnetmobile(Deep Learning Toolbox).

Yes

Yes

No

ResNet

ResNet-18, ResNet-50, and ResNet-101 convolutional neural networks. For the pretrained ResNet models, seeresnet50(Deep Learning Toolbox),resnet18(Deep Learning Toolbox), andresnet101(Deep Learning Toolbox).

The syntaxresnetXX('Weights','none')is not supported for code generation.

Yes

Yes

Yes

SegNet

Multi-class pixelwise segmentation network. For more information, seesegnetLayers(Computer Vision Toolbox).

Yes

Yes

No

SqueezeNet

Small deep neural network. For the pretrained SqueezeNet models, seesqueezenet(Deep Learning Toolbox).

The syntaxsqueezenet('Weights','none')is not supported for code generation.

Yes

Yes

Yes

VGG-16

VGG-16 convolutional neural network. For the pretrained VGG-16 model, seevgg16(Deep Learning Toolbox).

The syntaxvgg16('Weights','none')is not supported for code generation.

Yes

Yes

Yes

VGG-19

VGG-19 convolutional neural network. For the pretrained VGG-19 model, seevgg19(Deep Learning Toolbox).

The syntaxvgg19('Weights','none')is not supported for code generation.

Yes

Yes

Yes

Xception

Xception convolutional neural network. For the pretrained Xception model, seexception(Deep Learning Toolbox).

The syntaxxception('Weights','none')is not supported for code generation.

Yes

Yes

Yes

YOLO v2

You only look once version 2 convolutional neural network based object detector. For more information, seeyolov2Layers(Computer Vision Toolbox)

Yes

Yes

Yes

Supported Layers

The following layers are supported for code generation by GPU Coder for the target deep learning libraries specified in the table.

Input Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

imageInputLayer(Deep Learning Toolbox)

An image input layer inputs 2-D images to a network and applies data normalization.

Code generation does not support'Normalization'specified using a function handle.

Yes

Yes

Yes

sequenceInputLayer(Deep Learning Toolbox)

A sequence input layer inputs sequence data to a network.

The cuDNN library supports vector and 2-D image sequences. The TensorRT library support only vector input sequences.

For vector sequence inputs, the number of features must be a constant during code generation.

For image sequence inputs, the height, width, and the number of channels must be a constant during code generation.

Code generation does not support'Normalization'specified using a function handle.

Yes

Yes

No

Convolution and Fully Connected Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

convolution2dLayer(Deep Learning Toolbox)

A 2-D convolutional layer applies sliding convolutional filters to the input.

Yes

Yes

Yes

fullyConnectedLayer(Deep Learning Toolbox)

A fully connected layer multiplies the input by a weight matrix and then adds a bias vector.

Yes

Yes

No

groupedConvolution2dLayer(Deep Learning Toolbox)

A 2-D grouped convolutional layer separates the input channels into groups and applies sliding convolutional filters. Use grouped convolutional layers for channel-wise separable (also known as depth-wise separable) convolution.

Code generation for the ARM Mali GPU is not supported for a 2-D grouped convolution layer that has theNumGroupsproperty set as'channel-wise'or a value greater than two.

Yes

Yes

Yes

transposedConv2dLayer(Deep Learning Toolbox)

A transposed 2-D convolution layer upsamples feature maps.

Yes

Yes

Yes

Sequence Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

bilstmLayer(Deep Learning Toolbox)

A bidirectional LSTM (BiLSTM) layer learns bidirectional long-term dependencies between time steps of time series or sequence data. These dependencies can be useful when you want the network to learn from the complete time series at each time step.

For code generation, theStateActivationFunctionproperty must be set to'tanh'.

For code generation, theGateActivationFunctionproperty must be set to'sigmoid'.

Yes

Yes

No

flattenLayer(Deep Learning Toolbox)

A flatten layer collapses the spatial dimensions of the input into the channel dimension.

Yes

No

No

gruLayer(Deep Learning Toolbox)

A GRU layer learns dependencies between time steps in time series and sequence data.

Code generation supports only the'after-multiplication'and'recurrent-bias-after-multiplication'reset gate modes.

Yes

Yes

No

lstmLayer(Deep Learning Toolbox)

An LSTM layer learns long-term dependencies between time steps in time series and sequence data.

For code generation, theStateActivationFunctionproperty must be set to'tanh'.

For code generation, theGateActivationFunctionproperty must be set to'sigmoid'.

Yes

Yes

No

sequenceFoldingLayer(Deep Learning Toolbox)

A sequence folding layer converts a batch of image sequences to a batch of images. Use a sequence folding layer to perform convolution operations on time steps of image sequences independently.

Yes

No

No

sequenceInputLayer(Deep Learning Toolbox)

A sequence input layer inputs sequence data to a network.

The cuDNN library supports vector and 2-D image sequences. The TensorRT library support only vector input sequences.

For vector sequence inputs, the number of features must be a constant during code generation.

For image sequence inputs, the height, width, and the number of channels must be a constant during code generation.

Code generation does not support'Normalization'specified using a function handle.

Yes

Yes

No

sequenceUnfoldingLayer(Deep Learning Toolbox)

A sequence unfolding layer restores the sequence structure of the input data after sequence folding.

Yes

No

No

wordEmbeddingLayer(Text Analytics Toolbox)

A word embedding layer maps word indices to vectors.

Yes

Yes

No

Activation Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

clippedReluLayer(Deep Learning Toolbox)

A clipped ReLU layer performs a threshold operation, where any input value less than zero is set to zero and any value above theclipping ceilingis set to that clipping ceiling.

Yes

Yes

Yes

eluLayer(Deep Learning Toolbox)

An ELU activation layer performs the identity operation on positive inputs and an exponential nonlinearity on negative inputs.

Yes

Yes

No

leakyReluLayer(Deep Learning Toolbox)

A leaky ReLU layer performs a threshold operation, where any input value less than zero is multiplied by a fixed scalar.

Yes

Yes

Yes

reluLayer(Deep Learning Toolbox)

A ReLU layer performs a threshold operation to each element of the input, where any value less than zero is set to zero.

Yes

Yes

Yes

softplusLayer(Reinforcement Learning Toolbox)

ASoftplusLayeris a deep neural network layer that implements the softplus activationY= log(1 + eX), which ensures that the output is always positive

.

Yes

Yes

No

tanhLayer(Deep Learning Toolbox)

A hyperbolic tangent (tanh) activation layer applies the tanh function on the layer inputs.

Yes

Yes

Yes

Normalization, Dropout, and Cropping Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

batchNormalizationLayer(Deep Learning Toolbox)

A batch normalization layer normalizes each input channel across a mini-batch.

Yes

Yes

Yes

crop2dLayer(Deep Learning Toolbox)

A 2-D crop layer applies 2-D cropping to the input.

Yes

Yes

Yes

crossChannelNormalizationLayer(Deep Learning Toolbox)

A channel-wise local response (cross-channel) normalization layer carries out channel-wise normalization.

Yes

Yes

Yes

dropoutLayer(Deep Learning Toolbox)

A dropout layer randomly sets input elements to zero with a given probability.

Yes

Yes

Yes

scalingLayer(Reinforcement Learning Toolbox)

Scaling layer for actor or critic network.

For code generation, values for the'Scale'and'Bias'properties must have the same dimension.

Yes

Yes

Yes

Pooling and Unpooling Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

averagePooling2dLayer(Deep Learning Toolbox)

An average pooling layer performs down-sampling by dividing the input into rectangular pooling regions and computing the average values of each region.

Yes

Yes

Yes

globalAveragePooling2dLayer(Deep Learning Toolbox)

A global average pooling layer performs down-sampling by computing the mean of the height and width dimensions of the input.

Yes

Yes

Yes

globalMaxPooling2dLayer(Deep Learning Toolbox)

A global max pooling layer performs down-sampling by computing the maximum of the height and width dimensions of the input.

Yes

Yes

Yes

maxPooling2dLayer(Deep Learning Toolbox)

A max pooling layer performs down-sampling by dividing the input into rectangular pooling regions, and computing the maximum of each region.

Yes

Yes

Yes

maxUnpooling2dLayer(Deep Learning Toolbox)

马克斯unpooling层unpools最大的输出pooling layer.

Yes

Yes

No

Combination Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

additionLayer(Deep Learning Toolbox)

An addition layer adds inputs from multiple neural network layers element-wise.

Yes

Yes

Yes

concatenationLayer(Deep Learning Toolbox)

A concatenation layer takes inputs and concatenates them along a specified dimension.

Yes

Yes

No

depthConcatenationLayer(Deep Learning Toolbox)

A depth concatenation layer takes inputs that have the same height and width and concatenates them along the third dimension (the channel dimension).

Yes

Yes

Yes

Object Detection Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

anchorBoxLayer(Computer Vision Toolbox)

锚箱层商店锚箱的壮举ure map used in object detection networks.

Yes

Yes

Yes

focalLossLayer(Computer Vision Toolbox)

A focal loss layer predicts object classes using focal loss.

Yes

Yes

Yes

spaceToDepthLayer(Computer Vision Toolbox)

A space to depth layer permutes the spatial blocks of the input into the depth dimension. Use this layer when you need to combine feature maps of different size without discarding any feature data.

Yes

Yes

Yes

ssdMergeLayer(Computer Vision Toolbox)

An SSD merge layer merges the outputs of feature maps for subsequent regression and classification loss computation.

Yes

Yes

No

rcnnBoxRegressionLayer(Computer Vision Toolbox)

A box regression layer refines bounding box locations by using a smooth L1 loss function. Use this layer to create a Fast or Faster R-CNN object detection network.

Yes

Yes

Yes

rpnClassificationLayer(Computer Vision Toolbox)

A region proposal network (RPN) classification layer classifies image regions as eitherobjectorbackgroundby using a cross entropy loss function. Use this layer to create a Faster R-CNN object detection network.

Yes

Yes

Yes

YOLOv2OutputLayer(Computer Vision Toolbox)

Create output layer for YOLO v2 object detection network.

Yes

Yes

Yes

YOLOv2ReorgLayer(Computer Vision Toolbox)

Create reorganization layer for YOLO v2 object detection network.

Yes

Yes

Yes

YOLOv2TransformLayer(Computer Vision Toolbox)

Create transform layer for YOLO v2 object detection network.

Yes

Yes

Yes

Output Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

classificationLayer(Deep Learning Toolbox)

A classification layer computes the cross entropy loss for multi-class classification problems with mutually exclusive classes.

Yes

Yes

Yes

dicePixelClassificationLayer(Computer Vision Toolbox)

A Dice pixel classification layer provides a categorical label for each image pixel or voxel using generalized Dice loss.

Yes

Yes

Yes

focalLossLayer(Computer Vision Toolbox)

A focal loss layer predicts object classes using focal loss.

Yes

Yes

Yes

Output Layer(Deep Learning Toolbox)

All output layers including custom classification or regression output layers created by usingnnet.layer.ClassificationLayerornnet.layer.RegressionLayer.

For an example showing how to define a custom classification output layer and specify a loss function, seeDefine Custom Classification Output Layer(Deep Learning Toolbox).

For an example showing how to define a custom regression output layer and specify a loss function, seeDefine Custom Regression Output Layer(Deep Learning Toolbox).

Yes

Yes

Yes

pixelClassificationLayer(Computer Vision Toolbox)

A pixel classification layer provides a categorical label for each image pixel or voxel.

Yes

Yes

Yes

rcnnBoxRegressionLayer(Computer Vision Toolbox)

A box regression layer refines bounding box locations by using a smooth L1 loss function. Use this layer to create a Fast or Faster R-CNN object detection network.

Yes

Yes

Yes

regressionLayer(Deep Learning Toolbox)

A regression layer computes the half-mean-squared-error loss for regression problems.

Yes

Yes

Yes

rpnClassificationLayer(Computer Vision Toolbox)

A region proposal network (RPN) classification layer classifies image regions as eitherobjectorbackgroundby using a cross entropy loss function. Use this layer to create a Faster R-CNN object detection network.

Yes

Yes

Yes

sigmoidLayer(Deep Learning Toolbox)

A sigmoid layer applies a sigmoid function to the input.

Yes

Yes

Yes

softmaxLayer(Deep Learning Toolbox)

A softmax layer applies a softmax function to the input.

Yes

Yes

Yes

Keras and ONNX Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

nnet.keras.layer.FlattenCStyleLayer(Deep Learning Toolbox)

Flatten activations into 1-D assuming C-style (row-major) order.

Yes

Yes

Yes

nnet.keras.layer.GlobalAveragePooling2dLayer(Deep Learning Toolbox)

Global average pooling layer for spatial data.

Yes

Yes

Yes

nnet.keras.layer.SigmoidLayer(Deep Learning Toolbox)

Sigmoid activation layer.

Yes

Yes

Yes

nnet.keras.layer.TanhLayer(Deep Learning Toolbox)

Hyperbolic tangent activation layer.

Yes

Yes

Yes

nnet.keras.layer.ZeroPadding2dLayer(Deep Learning Toolbox)

Zero padding layer for 2-D input.

Yes

Yes

Yes

nnet.onnx.layer.ElementwiseAffineLayer(Deep Learning Toolbox)

Layer that performs element-wise scaling of the input followed by an addition.

Yes

Yes

Yes

nnet.onnx.layer.FlattenLayer(Deep Learning Toolbox)

Flattens the spatial dimensions of the input tensor to the channel dimensions.

Yes

Yes

Yes

nnet.onnx.layer.IdentityLayer(Deep Learning Toolbox)

Layer that implements ONNX identity operator.

Yes

Yes

Yes

Custom Layers

Layer Name Description cuDNN TensorRT ARM Compute Library for Mali GPU

Custom layers

Custom layers, with or without learnable parameters, that you define for your problem.

To learn how to define custom deep learning layers, seeDefine Custom Deep Learning Layers(Deep Learning Toolbox)andDefine Custom Deep Learning Layer for Code Generation(Deep Learning Toolbox).

For an example on how to generate code for a network with custom layers, seeCode Generation For Object Detection Using YOLO v3 Deep Learning.

The outputs of the custom layer must be fixed-size arrays.

Using'unified'as theMallocModecoder.gpuConfigrequires extra memory copies leading to slower performance. For custom layers, it is recommended to use'discrete'mode. For more information on GPU memory allocation, seeDiscrete and Managed Modes

cuDNN targets support both row-major and column-major code generation for custom layers. TensorRT targets support only column-major code generation.

Code generation for a sequence network containing custom layer and LSTM or GRU layer is not supported.

Code generation for a deep learning network with custom layer is not supported in Simulink®.

Yes

Yes

No

Supported Classes

The following classes are supported for code generation by GPU Coder for the target deep learning libraries specified in the table.

Name Description cuDNN TensorRT ARM Compute Library for Mali GPU
DAGNetwork(Deep Learning Toolbox)

Directed acyclic graph (DAG) network for deep learning

  • Only theactivations,predict, andclassifymethods are supported.

Yes

Yes

Yes

SeriesNetwork(Deep Learning Toolbox)

Series network for deep learning

  • Only theactivations,classify,predict,predictAndUpdateState, andresetStateobject functions are supported.

Yes

Yes

Yes

ssdObjectDetector(Computer Vision Toolbox)

Object to detect objects using the SSD-based detector.

  • Only thedetect(Computer Vision Toolbox)method of thessdObjectDetectoris supported for code generation.

  • Theroiargument to thedetectmethod must be a codegen constant (coder.const()) and a 1x4 vector.

  • Only theThreshold,SelectStrongest,MinSize,MaxSize, andMiniBatchSizeName-Value pairs are supported. All Name-Value pairs must be compile-time constants.

  • The channel and batch size of the input image must be fixed size.

  • Thelabelsoutput is returned as a categorical array.

  • In the generated code, the input is rescaled to the size of the input layer of the network. But the bounding box that thedetectmethod returns is in reference to the original input size.

  • The bounding boxes might not numerically match the simulation results.

Yes

Yes

No

yolov2ObjectDetector(Computer Vision Toolbox)

Detect objects using YOLO v2 object detector

  • Only thedetect(Computer Vision Toolbox)method of theyolov2ObjectDetectoris supported for code generation.

  • Theroiargument to thedetectmethod must be a codegen constant (coder.const()) and a 1x4 vector.

  • Only theThreshold,SelectStrongest,MinSize,MaxSize, andMiniBatchSizeName-Value pairs are supported.

  • The height, width, channel, and batch size of the input image must be fixed size.

  • The minimum batch size value passed to detect method must be fixed size.

  • The labels output is returned as a cell array of character vectors, such as {'car','bus'}.

Yes

Yes

Yes

See Also

Functions

Objects

Related Topics