Loading [MathJax]/jax/output/HTML-CSS/jax.js

Edge histogram based features

Press the button 'Toggle code' below to toggle code on and off for entire this presentation.

Figure 1: (left panel) A natural image (in this instance of the two creators/writers of the television show 'South Park' (this image is reproduced with permission of Jason Marck). (right panel) The edge-detected version of this image, where the bright yellow pixels indicate large edge content, still describes the scene very well (in the sense that we can still tell there are two people in the image) using only a fraction of the information contained in the original image. Note that edges have been colored yellow for visualization purposes only.
Figure 2: Visual information is processed in an area of the brain where each neuron detects in the observed scene edges of a specific orientation and width. It is thought that what we (and other mammals) "see" is a processed interpolation of these edge detected images.

Edge detection via convolution

To capture the total 'edge content' of an image in each direction, we

  • I. convolve it with the appropriate kernel
  • II. pass the results through a rectified linear unit (ReLU) to remove negative entries
  • III. sum up (the remaining positive) pixel values
fi=all pixelsmax(0,wix)i=1,,8

Why pass the convolution through ReLU?

  • Real images used in practice are significantly more complicated than these simplistic geometrical shapes and summarizing them using just eight features would be extremely ineffective.
  • To fix this issue, instead of computing the edge histogram features over the entire image, we break the image into relatively small patches, and compute features for each patch
fi,j=jth patchmax(0,wix)i=1,,8
  • This is called pooling.

Putting all pieces together: the complete feature extraction pipeline