How to add differential functions and other ops to SameDiff graph.
To get started with SameDiff, familiarize yourself with the autodiff
module of the ND4J API located here on GitHub.
For better or worse, SameDiff code is organized in just a few key places. For basic usage and testing of SameDiff the following modules are key. We'll discuss some of them in more detail in just a bit.
functions
: This module has the basic building blocks to build SameDiff variables and graphs.
execution
: has everything related to SameDiff graph execution.
gradcheck
: Utility functionality for checking SameDiff gradients, similar in structure to the respective tool in DL4J.
loss
: Loss functions for SameDiff
samediff
: Main SameDiff module to define, set up and run SameDiff operations and graphs.
functions
moduleSee the functions
module on GitHub.
The central abstraction of the functions
module is DifferentialFunction
, which underlies pretty much everything in SameDiff. Mathematically, what we're doing in SameDiff is build a directed acyclic graph whose nodes are differential functions, for which we can compute gradients. In that regard, DifferentialFunction
makes up a SameDiff graph on a fundamental level.
Note that each DifferentialFunction
comes with a SameDiff
instance. We'll discuss SameDiff
and this relationship later on. Also, while there's only few key abstractions, they're essentially used everywhere, so it's almost impossible to discuss SameDiff concepts separately. Eventually we'll get around to each part.
Each differential function comes with properties. In the simplest case, a differential function just has a name. Depending on the operation in question, you'll usually have many more properties (think strides or kernel sizes in convolutions). When we import computation graphs from other projects (TensorFlow, ONNX, etc.) these properties need to be mapped to the conventions we're using internally. The methods attributeAdaptersForFunction
, mappingsForFunction
, propertiesForFunction
and resolvePropertiesFromSameDiffBeforeExecution
are what you want to look at to get started.
Once properties are defined and properly mapped, you call initFromTensorFlow
and initFromOnnx
for TensorFlow and ONNX import, respectively. More on this later, when we discuss building SameDiff operations.
A differential function is executed on a list of inputs, using function properties, and produces one or more output variables. You have access to many helper functions to set or access these variables:
args()
: returns all input variables.
arg()
: returns the first input variable (the only one for unary operations).
larg()
and rarg()
: return the first and second (read "left" and "right") argument for binary operations
outputVariables()
: returns a list of all output variables. Depending on the operation, this may be computed dynamically. As we'll see later on, to get the result for ops with a single output, we'll call .outputVariables()[0]
.
Handling output variables is tricky and one of the pitfalls in using and extending SameDiff. For instance, implementing calculateOutputShape
for a differential function might be necessary, but if implemented incorrectly can lead to hard-to-debug failures. (Note that SameDiff will eventually call op execution in libnd4j
and dynamic custom ops either infer output shapes or need to be provided with the correct output shape.)
Automatic differentiation for a differential functions is implemented in a single method: doDiff
. Each operation has to provide an implementation of doDiff
. If you're implementing a SameDiff operation for a libnd4j
op x
and you're lucky to find x_bp
(as in "back-propagation") you can use that and your doDiff
implementation comes essentially for free.
You'll also see a diff
implementation that's used internally and calls doDiff
.
Importantly, each differential function has access to a factory, an instance of DifferentialFunctionFactory
, by calling f()
. More precisely, this will return the factory of the SameDiff instance the differential function has:
This is used in many places and gives you access to all differential functions currently registered in SameDiff. Think of this factory as a provider of operations. Here's an example of exposing sum
to the DifferentialFunctionFactory
:
We leave out the function arguments on purpose here. Note that all we do is redirect to the Sum
operation defined elsewhere in ND4J and then return the first output variable (of type SDVariable
, discussed in a second). Disregarding the implementation details for now, what this allows you to do is call f().sum(...)
from anywhere you have access to a differential function factory. For instance, when implementing a SameDiff op x
and you already have x_bp
in your function factory, you can override doDiff
for x
samediff
See the samediff
module on GitHub.
Not surprisingly, this is where the magic happens. This module has the core structures that SameDiff operates with. First, let's have a look at the variables that make up SameDiff operations.
SDVariable
(read SameDiff variable) extends DifferentialFunction
and is to SameDiff what INDArray
is to good old ND4J. In particular, SameDiff graphs operate on these variables and each individual operation takes in and spits out a list of SDVariable
. An SDVariable
comes with a name, is equipped with a SameDiff
instance, has shape information and knows how to initialize itself with an ND4J WeightInitScheme
. You'll also find a few helpers to set and get these properties.
One of the few things an SDVariable
can do that a DifferentialFunction
can't it evaluate its result and return an underlying INDArray
by calling eval()
. This will run SameDiff internally and retrieve the result. A similar getter is getArr()
which you can call at any point to get the current value of this variable. This functionality is used extensively in testing, to assert proper results. An SDVariable
also has access to its current gradient through gradient()
. Upon initialization there won't be any gradient, it will usually be computed at a later point.
Apart from these methods, SDVariable
also carries methods for concrete ops (and is in that regard a little similar to DifferentialFunctionFactory
). For instance, defining add
as follows:
allows you to call c = a.add(b)
on two SameDiff variables, the result of which can be accessed by c.eval()
.
The SameDiff
class is the main workhorse of the module and brings together most of the concepts discussed so far. A little unfortunately, the inverse is also true and SameDiff
instances are part of all other SameDiff module abstractions in some way or the other (which is why you've seen it many times already). Generally speaking, SameDiff
is the main entry point for automatic differentiation and you use it to define a symbolic graph that carries operations on SDVariable
s. Once built, a SameDiff graph can be run in a few ways, for instance exec()
and execAndEndResult()
.
Convince yourself that invoking SameDiff()
sets up a million things! Essentially, SameDiff
will collect and give you access (in terms of both getters and setters) to
All differential functions for the graph, with all their properties, which can be accessed in various ways (e.g. name or id).
All inputs and output information for said functions.
All function properties and how to map them, propertiesToResolve
and propertiesForFunction
are of particular note.
SameDiff
is also the place where you expose new operations to the SameDiff module. Essentially, you write a little wrapper for the respective operation in the DifferentialFunctionFactory
instance f()
. Here's an example for cross products:
At this point it might be a good idea to check out and run a few examples. SameDiff tests are a good source for that. Here's an example of how to multiply two SameDiff variables
This example is taken from SameDiffTests, one of the main test sources, in which you also find a few complete end-to-end examples.
The second place you find tests is in samediff repo directory. Whenever you add a new operation to SameDiff, add tests for the forward pass and gradient checks as well.
The third set of relevant tests is stored in imports and contains test for importing TensorFlow and ONNX graphs. On a side note, the resources for these import tests are generated in our TFOpsTests project.
We've seen how ND4J operations get picked up by DifferentialFunctionFactory
and SameDiff
to expose them to SameDiff at various levels. As for actually implementing these ops, you need to know a few things. In libnd4j you find two classes of operations, which are described here in detail. We'll show how to implement both op types.
All operations go here, and most of the time it's obvious where exactly to put the ops. Special attention goes to layers
, which is reserved for deep learning layer implementations (like Conv2D
). These higher-level ops are based on the concept of Modules, similar to modules in pytorch or layers in TensorFlow. These layer op implementation also provide a source of more involved op implementations.
Legacy (or XYZ) operations are the old breed of ND4J operations with a characteristic "xyz" signature. Here's how to implement cosine in ND4J by wrapping the cos
legacy op from libn4j: Cosine implementation. When it comes to SameDiff, the good thing about legacy ops is that they're already available in ND4J, but need to be augmented by SameDiff specific functionality to pass the muster. Since the cosine function does not have any properties, this implementation is straightforward. The parts that make this op SameDiff compliant are:
You specify SameDiff constructors here
You implement doDiff
here.
You specify a SameDiff opName
, a TensorFlow tensorflowName
and an ONNX onnxName
here.
If you look closely, this is only part of the truth, since Cos
extends BaseTransformOp
, which implements other SameDiff functionality. (Note that BaseTransformOp
is a BaseOp
, which extends DifferentialFunction
from earlier.) For instance, calculateOutputShape
is implemented there. If you want to implement a new transform, you can simply inherit from BaseTransformOp
, too. For other op types like reductions etc. there are op base classes available as well, meaning you only need to address the three bullet points above.
In the rare case you need to write a legacy op from scratch, you'll have to find the respective op number from libn4j, which can be found in legacy_ops.h
.
DynamicCustomOp
is the new kind of operation from libnd4j and all recent additions are implemented as such. This operation type in ND4J directly extends DifferentialFunction
.
Here's an example of the BatchToSpace
operation, which inherits from DynamicCustomOp
:
BatchToSpace is initialized with two properties, blocks
and crops
. Note how blocks
and crops
, which are both of integer type, get added to integer arguments for the operation by calling addIArgument
. For float arguments and other types, use addTArgument
instead.
The operation gets its own name and names for import,
and doDiff
is implemented.
The BatchToSpace operation is then integrated into DifferentialFunctionFactory
here, exposed to SameDiff
here and tested here.
The only thing BatchToSpace is currently missing is property mapping. We call the properties for this operation blocks
and crops
, but in ONNX or TensorFlow they might be called and stored quite differently. To look up the differences for mappings this correctly, see ops.proto
for TensorFlow and onnxops.json
for ONNX.
Let's look at another operation that does property mapping right, namely DynamicPartition
. This op has precisely one property, called numPartitions
in SameDiff. To map and use this property, you do the following:
Implement a little helper method called addArgs
that is used in the constructor of the op and in an import helper one-liner that we're discussing next. It's not necessary, but encouraged to do this and call it addArgs
consistently, for clarity.
Override initFromTensorFlow
method that maps properties for us using a TFGraphMapper
instance and adding arguments with addArgs
. Note that since ONNX does not support dynamic partitioning at the time of this writing (hence no onnxName
) there's also no initFromOnnx
method, which works pretty much the same way as initFromTensorFlow
.
For the TensorFlow import to work, we also need to override mappingsForFunction
. This example of a mapping is very simple, all it does is map TensorFlow's property name num_partititions
to our name numPartitions
.
Note that while DynamicPartition
has proper property mapping, it currently does not have a working doDiff
implementation.
As a last example, we show one that has a little more interesting property mapping setup, namely Dilation2D
. Not only has this op far more properties to map, as you can see in mappingsForFunction
, the properties also come with property values, as defined in attributeAdaptersForFunction
. We've chosen to show this op because it is one that has property mapping, but is neither exposed to DifferentialFunctionFactory
not SameDiff
.
Hence, the three DynamicCustomOp
examples shown each come with their own defects and represent examples of the work that has to be done for SameDiff. To summarize, to add a new SameDiff op you need to:
Create a new operation in ND4J that extends DifferentialFunction
. How exactly this implementation is set up depends on the
op generation (legacy vs. dynamic custom)
op type (transform, reduction, etc.)
Define an own op name, as well as TensorFlow and ONNX names.
Define necessary SameDiff constructors
Use addArgs
to add op arguments in a reusable way.
Expose the operation in DifferentialFunctionFactory
first and wrap it then in SameDiff
(or SDVariable
for variable methods).
Implement doDiff
for automatic differentiation.
Override mappingsForFunction
to map properties for TensorFlow and ONNX
If necessary, also provide an attribute adapter by overriding attributeAdaptersForFunction
.
Add import one-liners for TensorFlow and ONNX by adding initFromTensorFlow
and initFromOnnx
(using addArgs
).
Test, test, test.