Python API¶
This section includes information for using the pure Python API of bob.ip.optflow.hornschunck
.

bob.ip.optflow.hornschunck.
get_config
()[source]¶ Returns a string containing the configuration information.

class
bob.ip.optflow.hornschunck.
CentralGradient
¶ Bases:
object
Computes the spatiotemporal gradient using a 3term approximation
This class computes the spatiotemporal gradient using a 3term approximation composed of 2 separable kernels (one for the diference term and another one for the averaging term).
Constructor Documentation:
bob.ip.optflow.hornschunck.CentralGradient (difference, average, (height, width))
Constructor
We initialize with the shape of the images we need to treat and with the kernels to be applied. The shape is used by the internal buffers.
Parameters:
difference
: arraylike, 1D float64The kernel that contains the difference operation. Typically, this is[1, 0, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, 0, +1]
sliding operator, specify[+1, 0, 1]
. This kernel must have a shape = (3,).average
: arraylike, 1D float64The kernel that contains the spatial averaging operation. This kernel is typically[+1, +1, +1]
. This kernel must have a shape = (3,).(height, width)
: tuplethe height and width of images to be fed into the the gradient estimatorClass Members:

average
¶ numpy.ndarray
<– The kernel that contains the average operation (1D array of typefloat64
). Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).

difference
¶ numpy.ndarray
<– The kernel that contains the difference operation (1D array of typefloat64
). Typically, this is[1, 0, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, 0, +1]
sliding operator, specify[+1, 0, 1]
. This kernel must have a shape = (3,).

evaluate
(image1, image2, image3[, ex, ey, et]) → ex, ey, et¶ Evaluates the spatiotemporal gradient from the input image tripplet
Parameters:
image1, image2, image3
: arraylike (2D, float64)Sequence of images to evaluate the gradient from. All images should have the same shape, which should match that of this functor. The gradient is evaluated w.r.t. the image in the center of the tripplet.ex, ey, et
: array (2D, float64)The evaluated gradients in the horizontal, vertical and time directions (respectively) will be output in these variables, which should have dimensions matching those of this functor. If you don’t provide arrays forex
,ey
andet
, then they will be allocated internally and returned. You must either provide neitherex
,ey
andet
or all, otherwise an exception will be raised.Returns:
ex, ey, et
: array (2D, float64)The evaluated gradients are returned by this function. Each matrix will have a shape that matches the input images.

shape
¶ tuple
<– The shape preconfigured for this gradient estimator:(height, width)


class
bob.ip.optflow.hornschunck.
Flow
¶ Bases:
object
Estimates the Optical Flow between images.
This is a clone of the Vanilla Horn & Schunck method that uses a Sobel gradient estimator instead of the forward estimator used by the classical method. The Laplacian operator is also replaced with a more common implementation. The Sobel filter requires 3 images for the gradient estimation. Therefore, this implementation inputs 3 image sets instead of just 2. The flow is calculated w.r.t. central image.
For more details on the general technique from Horn & Schunck, see the module’s documentation.
Constructor Documentation:
bob.ip.optflow.hornschunck.Flow ((height, width))
Initializes the functor with the sizes of images to be treated.
Parameters:
(height, width)
: tuplethe height and width of images to be fed into the the flow estimatorClass Members:

estimate
(alpha, iterations, image1, image2, image3[, u, v]) → u, v¶ Estimates the optical flow leading to
image2
. This method will use leading imageimage1
and the after imageimage3
, to estimate the optical flow leading toimage2
. All input images should be 2D 64bit float arrays with the shape(height, width)
as specified in the construction of the object.Parameters:
alpha
: floatThe weighting factor between brightness constness and the field smoothness. According to original paper, should be more or less set to noise in estimating . In practice, many algorithms consider values around 200 a good default. The higher this number is, the more importance on smoothing you will be putting.iterations
: intNumber of iterations for which to minimize the flow errorimage1, image2, image3
: arraylike (2D, float64)Sequence of images to estimate the flow fromu, v
: array (2D, float64)The estimated flows in the horizontal and vertical directions (respectively) will be output in these variables, which should have dimensions matching those of this functor. If you don’t provide arrays foru
andv
, then they will be allocated internally and returned. You must either provide neitheru
andv
or both, otherwise an exception will be raised. Notice that, if you provideu
andv
which are nonzero, they will be taken as initial values for the error minimization. These arrays will be updated with the final value of the flow leading toimage2
.Returns:
u, v
: array (2D, float)The estimated flows in the horizontal and vertical directions (respectively).

eval_eb
(image1, image2, image3, u, v) → error¶ Calculates the brightness error () as defined in the paper:
Parameters:
image1, image2, image3
: arraylike (2D, float64)Sequence of images the flow was estimated withu, v
: arraylike (2D, float64)The estimated flows in the horizontal and vertical directions (respectively), which should have dimensions matching those of this functor.Returns:
error
: array (2D, float)The evaluated brightness error.

eval_ec2
(u, v) → error¶ Calculates the square of the smoothness error () by using the formula described in the paper: . Sets the input matrix with the discrete values.
Parameters:
u, v
: arraylike (2D, float64)The estimated flows in the horizontal and vertical directions (respectively), which should have dimensions matching those of this functor.Returns:
error
: array (2D, float)The square of the smoothness error.

shape
¶ tuple
<– The shape preconfigured for this flow estimator:(height, width)


class
bob.ip.optflow.hornschunck.
ForwardGradient
¶ Bases:
object
Computes the spatiotemporal gradient using a 2term approximation
This class computes the spatiotemporal gradient using a 2term approximation composed of 2 separable kernels (one for the diference term and another one for the averaging term).
Constructor Documentation:
bob.ip.optflow.hornschunck.ForwardGradient (difference, average, (height, width))
Constructor
We initialize with the shape of the images we need to treat and with the kernels to be applied. The shape is used by the internal buffers.
Parameters:
difference
: arraylike, 1D float64The kernel that contains the difference operation. Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).average
: arraylike, 1D float64The kernel that contains the spatial averaging operation. This kernel is typically[+1, +1]
. This kernel must have a shape = (2,).(height, width)
: tuplethe height and width of images to be fed into the the gradient estimatorClass Members:

average
¶ numpy.ndarray
<– The kernel that contains the average operation (1D array of typefloat64
). Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).

difference
¶ numpy.ndarray
<– The kernel that contains the difference operation (1D array of typefloat64
). Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).

evaluate
(image1, image2[, ex, ey, et]) → ex, ey, et¶ Evaluates the spatiotemporal gradient from the input image pair
Parameters:
image1, image2
: arraylike (2D, float64)Sequence of images to evaluate the gradient from. Both images should have the same shape, which should match that of this functor.ex, ey, et
: array (2D, float64)The evaluated gradients in the horizontal, vertical and time directions (respectively) will be output in these variables, which should have dimensions matching those of this functor. If you don’t provide arrays forex
,ey
andet
, then they will be allocated internally and returned. You must either provide neitherex
,ey
andet
or all, otherwise an exception will be raised.Returns:
ex, ey, et
: array (2D, float64)The evaluated gradients are returned by this function. Each matrix will have a shape that matches the input images.

shape
¶ tuple
<– The shape preconfigured for this gradient estimator:(height, width)


class
bob.ip.optflow.hornschunck.
HornAndSchunckGradient
¶ Bases:
bob.ip.optflow.hornschunck.ForwardGradient
Computes the spatiotemporal gradient using a 2term approximation
This class computes the spatiotemporal gradient using the same approximation as the one described by Horn & Schunck in the paper titled ‘Determining Optical Flow’, published in 1981, Artificial Intelligence, * Vol. 17, No. 13, pp. 185203.
This is equivalent to convolving the image sequence with the following (separate) kernels:
Constructor Documentation:
bob.ip.optflow.hornschunck.HornAndSchunckGradient ((height, width))
Constructor
We initialize with the shape of the images we need to treat. The shape is used by the internal buffers.
The difference kernel for this operator is fixed to . The averaging kernel is fixed to .
Parameters:
(height, width)
: tuplethe height and width of images to be fed into the the gradient estimatorClass Members:

average
¶ numpy.ndarray
<– The kernel that contains the average operation (1D array of typefloat64
). Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).

difference
¶ numpy.ndarray
<– The kernel that contains the difference operation (1D array of typefloat64
). Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).

evaluate
(image1, image2[, ex, ey, et]) → ex, ey, et¶ Evaluates the spatiotemporal gradient from the input image pair
Parameters:
image1, image2
: arraylike (2D, float64)Sequence of images to evaluate the gradient from. Both images should have the same shape, which should match that of this functor.ex, ey, et
: array (2D, float64)The evaluated gradients in the horizontal, vertical and time directions (respectively) will be output in these variables, which should have dimensions matching those of this functor. If you don’t provide arrays forex
,ey
andet
, then they will be allocated internally and returned. You must either provide neitherex
,ey
andet
or all, otherwise an exception will be raised.Returns:
ex, ey, et
: array (2D, float64)The evaluated gradients are returned by this function. Each matrix will have a shape that matches the input images.

shape
¶ tuple
<– The shape preconfigured for this gradient estimator:(height, width)


class
bob.ip.optflow.hornschunck.
IsotropicGradient
¶ Bases:
bob.ip.optflow.hornschunck.CentralGradient
Computes the spatiotemporal gradient using a Isotropic filter
This class computes the spatiotemporal gradient using a 3D isotropic filter. The gradients are calculated along the x, y and t directions. The Isotropic operator can be decomposed into 2 1D kernels that are applied in sequence. Considering and one can represent the operations like this:
This is equivalent to convolving the image sequence with the following (separate) kernels:
Constructor Documentation:
bob.ip.optflow.hornschunck.IsotropicGradient ((height, width))
Constructor
We initialize with the shape of the images we need to treat. The shape is used by the internal buffers.
The difference kernel for this operator is fixed to . The averaging kernel is fixed to .
Parameters:
(height, width)
: tuplethe height and width of images to be fed into the the gradient estimatorClass Members:

average
¶ numpy.ndarray
<– The kernel that contains the average operation (1D array of typefloat64
). Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).

difference
¶ numpy.ndarray
<– The kernel that contains the difference operation (1D array of typefloat64
). Typically, this is[1, 0, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, 0, +1]
sliding operator, specify[+1, 0, 1]
. This kernel must have a shape = (3,).

evaluate
(image1, image2, image3[, ex, ey, et]) → ex, ey, et¶ Evaluates the spatiotemporal gradient from the input image tripplet
Parameters:
image1, image2, image3
: arraylike (2D, float64)Sequence of images to evaluate the gradient from. All images should have the same shape, which should match that of this functor. The gradient is evaluated w.r.t. the image in the center of the tripplet.ex, ey, et
: array (2D, float64)The evaluated gradients in the horizontal, vertical and time directions (respectively) will be output in these variables, which should have dimensions matching those of this functor. If you don’t provide arrays forex
,ey
andet
, then they will be allocated internally and returned. You must either provide neitherex
,ey
andet
or all, otherwise an exception will be raised.Returns:
ex, ey, et
: array (2D, float64)The evaluated gradients are returned by this function. Each matrix will have a shape that matches the input images.

shape
¶ tuple
<– The shape preconfigured for this gradient estimator:(height, width)


class
bob.ip.optflow.hornschunck.
PrewittGradient
¶ Bases:
bob.ip.optflow.hornschunck.CentralGradient
Computes the spatiotemporal gradient using a Prewitt filter
This class computes the spatiotemporal gradient using a 3D prewitt filter. The gradients are calculated along the x, y and t directions. The Prewitt operator can be decomposed into 2 1D kernels that are applied in sequence. Considering and one can represent the operations like this:
This is equivalent to convolving the image sequence with the following (separate) kernels:
Constructor Documentation:
bob.ip.optflow.hornschunck.PrewittGradient ((height, width))
Constructor
We initialize with the shape of the images we need to treat. The shape is used by the internal buffers.
The difference kernel for this operator is fixed to . The averaging kernel is fixed to .
Parameters:
(height, width)
: tuplethe height and width of images to be fed into the the gradient estimatorClass Members:

average
¶ numpy.ndarray
<– The kernel that contains the average operation (1D array of typefloat64
). Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).

difference
¶ numpy.ndarray
<– The kernel that contains the difference operation (1D array of typefloat64
). Typically, this is[1, 0, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, 0, +1]
sliding operator, specify[+1, 0, 1]
. This kernel must have a shape = (3,).

evaluate
(image1, image2, image3[, ex, ey, et]) → ex, ey, et¶ Evaluates the spatiotemporal gradient from the input image tripplet
Parameters:
image1, image2, image3
: arraylike (2D, float64)Sequence of images to evaluate the gradient from. All images should have the same shape, which should match that of this functor. The gradient is evaluated w.r.t. the image in the center of the tripplet.ex, ey, et
: array (2D, float64)The evaluated gradients in the horizontal, vertical and time directions (respectively) will be output in these variables, which should have dimensions matching those of this functor. If you don’t provide arrays forex
,ey
andet
, then they will be allocated internally and returned. You must either provide neitherex
,ey
andet
or all, otherwise an exception will be raised.Returns:
ex, ey, et
: array (2D, float64)The evaluated gradients are returned by this function. Each matrix will have a shape that matches the input images.

shape
¶ tuple
<– The shape preconfigured for this gradient estimator:(height, width)


class
bob.ip.optflow.hornschunck.
SobelGradient
¶ Bases:
bob.ip.optflow.hornschunck.CentralGradient
Computes the spatiotemporal gradient using a Sobel filter
This class computes the spatiotemporal gradient using a 3D sobel filter. The gradients are calculated along the x, y and t directions. The Sobel operator can be decomposed into 2 1D kernels that are applied in sequence. Considering and one can represent the operations like this:
This is equivalent to convolving the image sequence with the following (separate) kernels:
Constructor Documentation:
bob.ip.optflow.hornschunck.SobelGradient ((height, width))
Constructor
We initialize with the shape of the images we need to treat. The shape is used by the internal buffers.
The difference kernel for this operator is fixed to . The averaging kernel is fixed to .
Parameters:
(height, width)
: tuplethe height and width of images to be fed into the the gradient estimatorClass Members:

average
¶ numpy.ndarray
<– The kernel that contains the average operation (1D array of typefloat64
). Typically, this is[1, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, +1]
sliding operator, specify[+1, 1]
. This kernel must have a shape = (2,).

difference
¶ numpy.ndarray
<– The kernel that contains the difference operation (1D array of typefloat64
). Typically, this is[1, 0, 1]
. Note the kernel is mirrored during the convolution operation. To obtain a[1, 0, +1]
sliding operator, specify[+1, 0, 1]
. This kernel must have a shape = (3,).

evaluate
(image1, image2, image3[, ex, ey, et]) → ex, ey, et¶ Evaluates the spatiotemporal gradient from the input image tripplet
Parameters:
image1, image2, image3
: arraylike (2D, float64)Sequence of images to evaluate the gradient from. All images should have the same shape, which should match that of this functor. The gradient is evaluated w.r.t. the image in the center of the tripplet.ex, ey, et
: array (2D, float64)The evaluated gradients in the horizontal, vertical and time directions (respectively) will be output in these variables, which should have dimensions matching those of this functor. If you don’t provide arrays forex
,ey
andet
, then they will be allocated internally and returned. You must either provide neitherex
,ey
andet
or all, otherwise an exception will be raised.Returns:
ex, ey, et
: array (2D, float64)The evaluated gradients are returned by this function. Each matrix will have a shape that matches the input images.

shape
¶ tuple
<– The shape preconfigured for this gradient estimator:(height, width)


class
bob.ip.optflow.hornschunck.
VanillaFlow
¶ Bases:
object
Estimates the Optical Flow between images.
Estimates the Optical Flow between two sequences of images (
image1
, the starting image andimage2
, the final image). It does this using the iterative method described by Horn & Schunck in the paper titled “Determining Optical Flow”, published in 1981, Artificial Intelligence, Vol. 17, No. 13, pp. 185203.The method constrains the calculation with two assertions that can be made on a natural sequence of images:
 For the same lighting conditions, the brightness () of the shapes in an image do not change and, therefore, the derivative of E w.r.t. time () equals zero.
 The relative velocities of adjancent points in an image varies smoothly. The smothness constraint is applied on the image data using the Laplacian operator.
It then approximates the calculation of conditions 1 and 2 above using a Taylor series expansion and ignoring terms with order greater or equal 2. This technique is also know as “Finite Differences” and is applied in other engineering fields such as Fluid Mechanics.
The problem is finally posed as an iterative process that simultaneouslyminimizes conditions 1 and 2 above. A weighting factor (  also sometimes referred as in some implementations) controls the relative importance of the two above conditions. The higher it gets, the smoother the field will be.
Note
OpenCV also has an implementation for H&S optical flow. It sets
This is the set of equations that are implemented:
Where:
 is the relative velocity in the direction
 is the relative velocity in the direction
 , and
 are partial derivatives of brightness in the , and
directions, estimated using finite differences based on the
input images,
i1
andi2
 laplacian estimates for x given equations in Section 8 of the paper
 laplacian estimates for y given equations in Section 8 of the paper
According to paper, should be more or less set to noise in estimating . In practice, many algorithms consider values around 200 a good default. The higher this number is, the more importance on smoothing you will be putting.
The initial conditions are set such that , except in the case where you provide them. For example, if you are analyzing a video stream, it is a good idea to use the previous estimate as the initial conditions.
Note
This is a dense flow estimator. The optical flow is computed for all pixels in the image.
Constructor Documentation:
bob.ip.optflow.hornschunck.VanillaFlow ((height, width))
Initializes the functor with the sizes of images to be treated.
Parameters:
(height, width)
: tuplethe height and width of images to be fed into the the flow estimatorClass Members:

estimate
(alpha, iterations, image1, image2[, u, v]) → u, v¶ Estimates the optical flow leading to
image2
. This method will use the leading imageimage1
, to estimate the optical flow leading toimage2
. All input images should be 2D 64bit float arrays with the shape(height, width)
as specified in the construction of the object.Parameters:
alpha
: floatThe weighting factor between brightness constness and the field smoothness. According to original paper, should be more or less set to noise in estimating . In practice, many algorithms consider values around 200 a good default. The higher this number is, the more importance on smoothing you will be putting.iterations
: intNumber of iterations for which to minimize the flow errorimage1, image2
: arraylike (2D, float64)Sequence of images to estimate the flow fromu, v
: array (2D, float64)The estimated flows in the horizontal and vertical directions (respectively) will be output in these variables, which should have dimensions matching those of this functor. If you don’t provide arrays foru
andv
, then they will be allocated internally and returned. You must either provide neitheru
andv
or both, otherwise an exception will be raised. Notice that, if you provideu
andv
which are nonzero, they will be taken as initial values for the error minimization. These arrays will be updated with the final value of the flow leading toimage2
.Returns:
u, v
: array (2D, float)The estimated flows in the horizontal and vertical directions (respectively).

eval_eb
(image1, image2, u, v) → error¶ Calculates the brightness error () as defined in the paper:
Parameters:
image1, image2
: arraylike (2D, float64)Sequence of images the flow was estimated withu, v
: arraylike (2D, float64)The estimated flows in the horizontal and vertical directions (respectively), which should have dimensions matching those of this functor.Returns:
error
: array (2D, float)The evaluated brightness error.

eval_ec2
(u, v) → error¶ Calculates the square of the smoothness error () by using the formula described in the paper: . Sets the input matrix with the discrete values.
Parameters:
u, v
: arraylike (2D, float64)The estimated flows in the horizontal and vertical directions (respectively), which should have dimensions matching those of this functor.Returns:
error
: array (2D, float)The square of the smoothness error.

shape
¶ tuple
<– The shape preconfigured for this flow estimator:(height, width)

bob.ip.optflow.hornschunck.
flow_error
(image1, image2, u, v) → E¶ Computes the generalized flow error between two images.
This function calculates the flow error between a pair of images:
Parameters:
image1, image2
: arraylike (2D, float64)Sequence of images the flow was estimated withu, v
: arraylike (2D, float64)The estimated flows in the horizontal and vertical directions (respectively), which should have dimensions matching those ofimage1
andimage2
.Returns:
E
: array (2D, float)The estimated flow error E.

bob.ip.optflow.hornschunck.
laplacian_avg_hs
(input) → output¶ Filters the input image using the Laplacian (averaging) operator.
An approximation to the Laplacian operator. Using the following (nonseparable) kernel:
This is the one used on the Horn & Schunck paper. To calculate the value we must remove the central mean and multiply by , yielding:
Note
You will get the wrong results if you use the Laplacian kernel directly.
Parameters:
input
: arraylike (2D, float64)The 2D array to which you’d like to apply the laplacian operator.Returns:
output
: array (2D, float)The result of applying the laplacian operator oninput
.

bob.ip.optflow.hornschunck.
laplacian_avg_hs_opencv
(input) → output¶ Filters the input image using the Laplacian (averaging) operator.
An approximation to the Laplacian operator. Using the following (nonseparable) kernel:
This is used as Laplacian operator in OpenCV. To calculate the value we must remove the central mean and multiply by , yielding:
Note
You will get the wrong results if you use the Laplacian kernel directly.
Parameters:
input
: arraylike (2D, float64)The 2D array to which you’d like to apply the laplacian operator.Returns:
output
: array (2D, float)The result of applying the laplacian operator oninput
.