[Ioannis mentioned another edge detector](http://www.parallelpoints.com/node/46#comment-132) he’s fond of using – so here’s a comparison.
To summarise, his method involves shifting the image 1 pixel to the right, and subtracting it from the original image. And then doing the same but shifting one pixel down instead.
This is equivalent to convolving with
[1 0 -1]
[2 0 -2]
[1 0 -1]
There’s two differences that are worth commenting on:
I’ve begun a series of articles showing the stages of development of an image processing system. The aim is to go through from describing algorithms at a very high level in Scilab, making tradeoffs in implementation ideas, before describing the system in VHDL and comparing the results.
[Part 1 – edge detection](/node/46) is now up…
# Image processing
This is the first in a group of posts, during which I’m going to work through the development of an image processing system. I’m going to start at a very high level working with some classic image processing techniques. The plan as I write this is to work down to a complete implemenation on an FPGA. In the process I’m going to demonstrate how to link the high-level descriptions to check the performance of the low-level FPGA code.