Deeplearning4j
Community ForumND4J JavadocDL4J Javadoc
EN 1.0.0-beta7
EN 1.0.0-beta7
  • Eclipse DeepLearning4J
  • Getting Started
    • Quickstart
      • Untitled
    • Tutorials
      • Quickstart with MNIST
      • MultiLayerNetwork And ComputationGraph
      • Logistic Regression
      • Built-in Data Iterators
      • Feed Forward Networks
      • Basic Autoencoder
      • Advanced Autoencoder
      • Convolutional Networks
      • Recurrent Networks
      • Early Stopping
      • Layers and Preprocessors
      • Hyperparameter Optimization
      • Using Multiple GPUs
      • Clinical Time Series LSTM
      • Sea Temperature Convolutional LSTM
      • Sea Temperature Convolutional LSTM 2
      • Instacart Multitask Example
      • Instacart Single Task Example
      • Cloud Detection Example
    • Core Concepts
    • Cheat Sheet
    • Examples Tour
    • Deep Learning Beginners
    • Build from Source
    • Contribute
      • Eclipse Contributors
    • Benchmark Guide
    • About
    • Release Notes
  • Configuration
    • Backends
      • CPU and AVX
      • cuDNN
      • Performance Issues
    • Memory Management
      • Memory Workspaces
    • Snapshots
    • Maven
    • SBT, Gradle, & Others
  • Models
    • Autoencoders
    • Multilayer Network
    • Computation Graph
    • Convolutional Neural Network
    • Recurrent Neural Network
    • Layers
    • Vertices
    • Iterators
    • Listeners
    • Custom Layers
    • Model Persistence
    • Activations
    • Updaters
  • Model Zoo
    • Overview
    • Zoo Models
  • ND4J
    • Overview
    • Quickstart
    • Basics
    • Elementwise Operations
    • Matrix Manipulation
    • Syntax
    • Tensors
  • SAMEDIFF
    • Importing TensorFlow models
    • Variables
    • Ops
    • Adding Ops
  • ND4J & SameDiff Ops
    • Overview
    • Bitwise
    • Linalg
    • Math
    • Random
    • BaseOps
    • CNN
    • Image
    • Loss
    • NN
    • RNN
  • Tuning & Training
    • Evaluation
    • Visualization
    • Trouble Shooting
    • Early Stopping
    • t-SNE Visualization
    • Transfer Learning
  • Keras Import
    • Overview
    • Get Started
    • Supported Features
      • Activations
      • Losses
      • Regularizers
      • Initializers
      • Constraints
      • Optimizers
    • Functional Model
    • Sequential Model
    • Custom Layers
    • API Reference
      • Core Layers
      • Convolutional Layers
      • Embedding Layers
      • Local Layers
      • Noise Layers
      • Normalization Layers
      • Pooling Layers
      • Recurrent Layers
      • Wrapper Layers
      • Advanced Activations
  • DISTRIBUTED DEEP LEARNING
    • Introduction/Getting Started
    • Technical Explanation
    • Spark Guide
    • Spark Data Pipelines Guide
    • API Reference
    • Parameter Server
  • Arbiter
    • Overview
    • Layer Spaces
    • Parameter Spaces
  • Datavec
    • Overview
    • Records
    • Reductions
    • Schema
    • Serialization
    • Transforms
    • Analysis
    • Readers
    • Conditions
    • Executors
    • Filters
    • Operations
    • Normalization
    • Visualization
  • Language Processing
    • Overview
    • Word2Vec
    • Doc2Vec
    • Sentence Iteration
    • Tokenization
    • Vocabulary Cache
  • Mobile (Android)
    • Setup
    • Tutorial: First Steps
    • Tutorial: Classifier
    • Tutorial: Image Classifier
    • FAQ
    • Press
    • Support
    • Why Deep Learning?
Powered by GitBook
On this page
  • Available layers
  • Convolution1D
  • Convolution2D
  • Convolution3D
  • Deconvolution2D
  • Cropping1D
  • Cropping2D

Was this helpful?

Edit on Git
Export as PDF
  1. Models

Convolutional Neural Network

Also known as CNN.

PreviousComputation GraphNextRecurrent Neural Network

Last updated 5 years ago

Was this helpful?

Available layers

Convolution1D

1D convolution layer. Expects input activations of shape [minibatch,channels,sequenceLength]

Convolution2D

2D convolution layer

Convolution3D

3D convolution layer configuration

hasBias

public boolean hasBias() 

An optional dataFormat: “NDHWC” or “NCDHW”. Defaults to “NCDHW”. The data format of the input and output data. For “NCDHW” (also known as ‘channels first’ format), the data storage order is: [batchSize, inputChannels, inputDepth, inputHeight, inputWidth]. For “NDHWC” (‘channels last’ format), the data is stored in the order of: [batchSize, inputDepth, inputHeight, inputWidth, inputChannels].

kernelSize

public Builder kernelSize(int... kernelSize) 

The data format for input and output activations. NCDHW: activations (in/out) should have shape [minibatch, channels, depth, height, width] NDHWC: activations (in/out) should have shape [minibatch, depth, height, width, channels]

stride

public Builder stride(int... stride) 

Set stride size for 3D convolutions in (depth, height, width) order

  • param stride kernel size

  • return 3D convolution layer builder

padding

public Builder padding(int... padding) 

Set padding size for 3D convolutions in (depth, height, width) order

  • param padding kernel size

  • return 3D convolution layer builder

dilation

public Builder dilation(int... dilation) 

Set dilation size for 3D convolutions in (depth, height, width) order

  • param dilation kernel size

  • return 3D convolution layer builder

dataFormat

public Builder dataFormat(DataFormat dataFormat) 

The data format for input and output activations. NCDHW: activations (in/out) should have shape [minibatch, channels, depth, height, width] NDHWC: activations (in/out) should have shape [minibatch, depth, height, width, channels]

  • param dataFormat Data format to use for activations

setKernelSize

public void setKernelSize(int... kernelSize) 

Set kernel size for 3D convolutions in (depth, height, width) order

  • param kernelSize kernel size

setStride

public void setStride(int... stride) 

Set stride size for 3D convolutions in (depth, height, width) order

  • param stride kernel size

setPadding

public void setPadding(int... padding) 

Set padding size for 3D convolutions in (depth, height, width) order

  • param padding kernel size

setDilation

public void setDilation(int... dilation) 

Set dilation size for 3D convolutions in (depth, height, width) order

  • param dilation kernel size

Deconvolution2D

2D deconvolution layer configuration

Deconvolutions are also known as transpose convolutions or fractionally strided convolutions. In essence, deconvolutions swap forward and backward pass with regular 2D convolutions.

hasBias

public boolean hasBias() 

Deconvolution2D layer nIn in the input layer is the number of channels nOut is the number of filters to be used in the net or in other words the channels The builder specifies the filter/kernel size, the stride and padding The pooling layer takes the kernel size

convolutionMode

public Builder convolutionMode(ConvolutionMode convolutionMode) 

Set the convolution mode for the Convolution layer. See {- link ConvolutionMode} for more details

  • param convolutionMode Convolution mode for layer

kernelSize

public Builder kernelSize(int... kernelSize) 

Size of the convolution rows/columns

  • param kernelSize the height and width of the kernel

Cropping1D

Cropping layer for convolutional (1d) neural networks. Allows cropping to be done separately for top/bottom

getOutputType

public InputType getOutputType(int layerIndex, InputType inputType) 
  • param cropTopBottom Amount of cropping to apply to both the top and the bottom of the input activations

setCropping

public void setCropping(int... cropping) 

Cropping amount for top/bottom (in that order). Must be length 1 or 2 array.

build

public Cropping1D build() 
  • param cropping Cropping amount for top/bottom (in that order). Must be length 1 or 2 array.

Cropping2D

Cropping layer for convolutional (2d) neural networks. Allows cropping to be done separately for top/bottom/left/right

getOutputType

public InputType getOutputType(int layerIndex, InputType inputType) 
  • param cropTopBottom Amount of cropping to apply to both the top and the bottom of the input activations

  • param cropLeftRight Amount of cropping to apply to both the left and the right of the input activations

setCropping

public void setCropping(int... cropping) 

Cropping amount for top/bottom/left/right (in that order). A length 4 array.

build

public Cropping2D build() 
  • param cropping Cropping amount for top/bottom/left/right (in that order). Must be length 4 array.

Cropping3D

Cropping layer for convolutional (3d) neural networks. Allows cropping to be done separately for upper and lower bounds of depth, height and width dimensions.

getOutputType

public InputType getOutputType(int layerIndex, InputType inputType) 
  • param cropDepth Amount of cropping to apply to both depth boundaries of the input activations

  • param cropHeight Amount of cropping to apply to both height boundaries of the input activations

  • param cropWidth Amount of cropping to apply to both width boundaries of the input activations

setCropping

public void setCropping(int... cropping) 

Cropping amount, a length 6 array, i.e. crop left depth, crop right depth, crop left height, crop right height, crop left width, crop right width

build

public Cropping3D build() 
  • param cropping Cropping amount, must be length 3 or 6 array, i.e. either crop depth, crop height, crop width or crop left depth, crop right depth, crop left height, crop right height, crop left width, crop right width

See the paper by Matt Zeiler for details:

For an intuitive guide to convolution arithmetic and shapes, see:

[source]
[source]
[source]
[source]
http://www.matthewzeiler.com/wp-content/uploads/2017/07/cvpr2010.pdf
https://arxiv.org/abs/1603.07285v1
[source]
[source]
[source]