.. _Neighborhood:
*************************************
Neighborhood (Spatial) Processing
*************************************
.. topic:: Reading Assignment
Please read section 12.5 of [RVC]_ and chapter 10 of [PIVP]_.
.. topic:: Video Resources
Peter Corke has a set of short `videos on this topic
`_.
This topic is usually called something like neighborhood processing,
spatial filtering, or something similar because the value of each output
pixel is determined by the corresponding pixel in the input image and the
neighboring pixels to it.
Convolution and Correlation
============================
Linear filtering of images apply a sum of products between pixels in a
neighborhood and a mask (also called a kernel) of values.
.. figure:: sum_prod.png
:align: center
:width: 70%
Figure 10.1 from [PIVP]_. The value of a pixel in the output image is the
sum of products between a neighborhood and a mask.
In signal processing in the time-domain, we can model how filters affect a
signal with a process called **convolution**. In convolution, we *fold* over
the signal and then at each time step shift the signal through the impulse
response of the filter. The output for each time step is the sum of the
products of the signal and filter response where they overlay each other.
Here is a simulated example of convolution in the time domain.
::
signal = [0 1 1 0.75 0.5 0 0 0 0 0];
filter = [0.75 0.5 0.25 0 0 0 0 0 0 0];
time = 0:9;
out = zeros(1,10);
for t = 1:10
for j=0:9
if t-j > 0
out(t) = out(t) + signal(t-j)*filter(j+1);
end
end
end
figure
subplot(3,1,1), stem(time, filter), title('Filter Response')
subplot(3,1,2), stem(time, signal), title('Input Signal')
subplot(3,1,3), stem(time, out), title('Filtered Signal')
.. figure:: time_convolve.png
:align: center
Example of how convolution models time-domain filtering
In the two dimensional spatial domain of images, we model linear
neighborhood filters with convolution when the filter mask is not symmetric
(mostly for edge detection). When the mask is symmetric, then convolution
and correlation yield the same result. **Correlation** is a sum of products
like convolution, but does not require folding the image or mask over.
Spatial domain 2D Convolution:
.. math:: \boldsymbol{\mathrm{O}}\left[u,v\right]=\sum_{(i,j)\in \mathcal{W}}
{\boldsymbol{\mathrm{I}}\left[u-i,v-j\right]\boldsymbol{\mathrm{K}}
\left[i,j\right],\forall (u,v)\in \boldsymbol{\mathrm{I}}} \\
i\in \left[-h\cdots h\right],\ j\in [-h\cdots h]
Spatial domain 2D Correlation:
.. math:: \boldsymbol{\mathrm{O}}\left[u,v\right]=\sum_{\left(i,j\right)\in
\mathcal{W}}{\mathrm{I}\left[u+i,v+j\right]\mathrm{K}\left[i,j\right],
\forall \left(u,v\right)\in \mathrm{I}} \\
i\in \left[-h\cdots h\right],\ j\in [-h\cdots h]
.. topic:: Properties of Convolution
.. describe:: commutative
:math:`A * B = B * A`
.. describe:: associative
:math:`A * B * C = (A * B) * C = A * (B * C)`
.. describe:: distributive
:math:`A * (B + C) = A * B + A * C`
.. describe:: linear
:math:`A * (\alpha B) = \alpha (A * B)`
The IPT function for applying a filter mask to an image is ``imfilter``.
We will learn about different options to ``imfilter`` in the tutorials.
The MVTB function for filtering images is ``iconvolve``
Border Geometry
-----------------
Look up the boundary options in the MATLAB documentation for the ``imfilter``
and ``iconvolve`` functions.
.. figure:: borders.png
:width: 50%
:align: center
Figure 10.4 from [PIVP]_.
Common Image Processing Filters
--------------------------------
So as to not change the brightness of an image, we usually form filter masks
such that the elements sum to one.
The IPT has a ``fspecial`` function that returns common filter masks.
The MVTB also has some functions that return filter masks.
The **average** or mean filter smooths or blurs an image.
Note that the average filter is not symetric with respect to rotation.
::
>> K = fspecial('average', 3)
K =
0.1111 0.1111 0.1111
0.1111 0.1111 0.1111
0.1111 0.1111 0.1111
>> K = ones(3)/3^2
K =
0.1111 0.1111 0.1111
0.1111 0.1111 0.1111
0.1111 0.1111 0.1111
The 'disk' option to ``fspecial`` and the ``kcircle`` function provide circular
averaging filters. A circular averaging filter is symmetric with respect to
rotation.
The **Gaussian** filter also smooths an image, but generally yields better
results than the average filters because it has smoother transitions between
pixels. It is based on the Gaussian function that is also used to describe a
normal probability distribution. The spread of the filter is defined by
standard deviation (:math:`\sigma`). Read the documentation in MATLAB for the
'gaussian' option to `fspecial` and the `kgauss` function. The Gaussian filter
is symmetric with respect to rotation and coefficients fall off to (almost)
zero at the kernel's edges.
.. math:: h(x, y) = e^{\frac{-(x^2 + y^2)}{2\sigma^2}}
The Laplacian
---------------
The Laplacian operator, is described for 2-D discrete valued functions, such
as an image, as a high-pass filter. It is the sum of the second partial
derivatives. While, not exactly *edge detection*, it places emphasis on the
edges of objects in an image.
.. math:: \nabla^2 f(x,y)
= \frac{\partial^2 f(x,y)}{\partial x^2}
+ \frac{\partial^2 f(x,y)}{\partial y^2}
Let's take a closer look at the second partial derivative term with respect
to :math:`x`. The first derivative is the difference in adjacent elements.
.. math:: \frac{\partial f(x, y) }{\partial x} = f'_x = f(x, y) - f(x - 1, y)
Or, if we want to look ahead of :math:`x`, instead of behind it,
it could also be:
.. math:: \frac{\partial f(x, y) }{\partial x} = f'_x = f(x + 1, y) - f(x, y)
Similarly, the second derivative is calculated from the difference between
adjacent first derivatives.
.. math:: \frac{\partial^2 f(x,y) }{\partial x^2} = f''_x(x, y)
= f'_x(x, y) - f'_x(x - 1, y)
Or:
.. math:: \frac{\partial^2 f(x,y) }{\partial x^2} = f''_x(x, y)
= f'_x(x + 1, y) - f'_x(x, y)
To be centered about :math:`x`, we will use the former expression for first
derivatives and the later expression for the second derivative.
.. math::
\begin{array}{ll} f''_x(x, y) & = f'_x(x + 1, y) - f'_x(x, y) \\
& = f(x + 1, y) - f(x, y) - f(x, y) + f(x - 1, y) \\
& = f(x + 1, y) + f(x - 1, y) - 2\,f(x, y)
\end{array}
The above equation is equation (10.13) from the text book. Equation (10.14)
with respect to the :math:`y` values follows likewise.
.. math::
\begin{array}{ll} f''_y(x, y) & = f'_y(x, y + 1) - f'_y(x, y) \\
& = f(x, y + 1) - f(x, y) - f(x, y) + f(x, y - 1) \\
& = f(x, y + 1) + f(x, y - 1) - 2\,f(x, y)
\end{array}
.. math:: \nabla^2 f(x,y)
= f(x + 1, y) + f(x - 1, y) +
f(x, y + 1) + f(x, y - 1) - 4\,f(x, y)
The convolution mask for the Laplacian is:
.. math:: \left[ \begin{array}{rrr}
0 & -1 & 0 \\
-1 & 4 & -1 \\
0 & -1 & 0
\end{array} \right]
Note that the signs of the coefficients in the convolution mask might be
reversed depending on the need.
Laplacian of Gaussian
-----------------------
If the edges in an image are first blurred, the Laplacian filter clearly
shows edges. This could be accomplished by two sequential filters, but
because of the associative property of convolution, we can used a single
filter that is the convolution of Laplacian and Gaussian filters, called
Laplacian of Gaussian or simply *LoG*.
::
% Demo to show how the Laplacian of Gaussian (log) filter works
% Make a simulated image showing half black and half white.
I = [zeros(256, 128) ones(256, 128)];
% Blurr the edge with a Gaussian filter
h = fspecial('gaussian', 10, 2);
Iblurr = imfilter(I, h);
% 1st and 2nd derivatives along u
deriv1 = [0 0 0; -1 1 0; 0 0 0];
deriv1_2 = [0 0 0; 0 -1 1; 0 0 0];
I1stD = imfilter(Iblurr, deriv1);
I2ndD = imfilter(I1stD, deriv1_2);
% Laplacian of Gaussian filter
hlog = fspecial('log', 10, 2);
Ilog = imfilter(I, hlog);
% plots
figure % images
subplot(2,3,1), imshow(I), title('original')
subplot(2,3,2), imshow(Iblurr), title('blurred')
subplot(2,3,3), imshow(I1stD, []), title('1st Derivative')
subplot(2,3,4), imshow(I2ndD, []), title('2nd Derivative')
subplot(2,3,5), imshow(Ilog, []), title('LoG')
figure % a horizontal line
subplot(2,3,1), plot(110:140, I(128,110:140)), title('original')
subplot(2,3,2), plot(110:140, Iblurr(128,110:140)), title('blurred')
subplot(2,3,3), plot(110:140, I1stD(128,110:140)), title('1st Derivative')
subplot(2,3,4), plot(110:140, I2ndD(128,110:140)), title('2nd Derivative')
subplot(2,3,5), plot(110:140, Ilog(128,110:140)), title('LoG')
Unsharp Filter
--------------
The oddly named unsharp filter is a high-pass filter that can make fine
detail in an image more noticeable. Its application is more related to image
processing than machine vision, but it is interesting enough to warrant brief
consideration.
The algorithm consists of computing the subtraction between the input image and
a blurred (low-pass filtered) version of the input image with the goal of
increasing the amount of high-frequency (fine) detail by reducing the
importance of its low-frequency contents. The "unsharp" of the name derives
from the fact that the technique uses a blurred, or "unsharp", negative image
to create a mask of the original image. The unsharped mask is then combined
with the positive (original) image, creating an image that is less blurry than
the original. The technique was first used and named with film photography. The
`Wikipedia page `_ shows some
diagrams that illustrate how the steps change an image.
The 'unsharp' option to the ``fspecial`` function provides a digital filter
mask for this affect. We will briefly demonstrate it in a tutorial.
Median Filter
=============
Images with *salt and pepper* noise, which is characterized by random white
and black pixels in the image, can benefit from the median filter. It
sets the target pixel equal to the median value of the neighborhood pixels.
The median filter is a *nonlinear* operation.
.. **Contents:**
.. .. toctree::
:maxdepth: 2