ZeePedia

Digital Image Definitions:COMMON VALUES, Types of operations, VIDEO PARAMETERS

<< Introduction
Tools:CONVOLUTION, FOURIER TRANSFORMS, Circularly symmetric signals >>
img
...Image Processing Fundamentals
regions­of­interest, ROIs, or simply regions. This concept reflects the fact that
images frequently contain collections of objects each of which can be the basis for a
region. In a sophisticated image processing system it should be possible to apply
specific image processing operations to selected regions. Thus one part of an image
(region) might be processed to suppress motion blur while another part might be
processed to improve color rendition.
The amplitudes of a given image will almost always be either real numbers or
integer numbers. The latter is usually a result of a quantization process that converts
a continuous range (say, between 0 and 100%) to a discrete number of levels. In
certain image-forming processes, however, the signal may involve photon counting
which implies that the amplitude would be inherently quantized. In other image
forming procedures, such as magnetic resonance imaging, the direct physical
measurement yields a complex number in the form of a real magnitude and a real
phase. For the remainder of this book we will consider amplitudes as reals or
integers unless otherwise indicated.
2.
Digital Image Definitions
A digital image a[m,n] described in a 2D discrete space is derived from an analog
image a(x,y) in a 2D continuous space through a sampling process that is
frequently referred to as digitization. The mathematics of that sampling process will
be described in Section 5. For now we will look at some basic definitions
associated with the digital image. The effect of digitization is shown in Figure 1.
The 2D continuous image a(x,y) is divided into N rows and M columns. The
intersection of a row and a column is termed a pixel. The value assigned to the
integer coordinates [m,n] with {m=0,1,2,...,M­1} and {n=0,1,2,...,N­1} is a[m,n].
In fact, in most cases a(x,y)--which we might consider to be the physical signal
that impinges on the face of a 2D sensor--is actually a function of many variables
including depth (z), color (λ), and time (t). Unless otherwise stated, we will
consider the case of 2D, monochromatic, static images in this chapter.
2
img
...Image Processing Fundamentals
Columns
Value = a(x, y, z, λ, t)
Figure 1: Digitization of a continuous image. The pixel at coordinates
[m=10, n=3] has the integer brightness value 110.
The image shown in Figure 1 has been divided into N = 16 rows and M = 16
columns. The value assigned to every pixel is the average brightness in the pixel
rounded to the nearest integer value. The process of representing the amplitude of
the 2D signal at a given coordinate as an integer value with L different gray levels is
usually referred to as amplitude quantization or simply quantization.
2.1 COMMON VALUES
There are standard values for the various parameters encountered in digital image
processing. These values can be caused by video standards, by algorithmic
requirements, or by the desire to keep digital circuitry simple. Table 1 gives some
commonly encountered values.
Parameter
Symbol
Typical values
N
Rows
256,512,525,625,1024,1035
M
Columns
256,512,768,1024,1320
L
Gray Levels
2,64,256,1024,4096,16384
Table 1: Common values of digital image parameters
Quite frequently we see cases of M=N=2K where {K = 8,9,10}. This can be
motivated by digital circuitry or by the use of certain algorithms such as the (fast)
Fourier transform (see Section 3.3).
3
img
...Image Processing Fundamentals
The number of distinct gray levels is usually a power of 2, that is, L=2B where B is
the number of bits in the binary representation of the brightness levels. When B >1
we speak of a gray-level image; when B =1 we speak of a binary image. In a binary
image there are just two gray levels which can be referred to, for example, as
"black" and "white" or "0" and "1".
2.2 CHARACTERISTICS OF IMAGE OPERATIONS
There is a variety of ways to classify and characterize image operations. The reason
for doing so is to understand what type of results we might expect to achieve with a
given type of operation or what might be the computational burden associated with
a given operation.
2.2.1 Types of operations
The types of operations that can be applied to digital images to transform an input
image a[m,n] into an output image b[m,n] (or another representation) can be
classified into three categories as shown in Table 2.
Operation
Characterization
Generic
Complexity/Pixel
· Point
constant
­ the output value at a specific coordinate is dependent
only on the input value at that same coordinate.
P2
· Local
­ the output value at a specific coordinate is dependent on
the input values in the neighborhood of that same
coordinate.
N2
· Global
­ the output value at a specific coordinate is dependent on
all the values in the input image.
Table 2: Types of image operations. Image size = N × N; neighborhood size
= P × P . Note that the complexity is specified in operations per pixel.
This is shown graphically in Figure 2.
a
b
a
b
Point
Local
a
Global
b
= [m=mo, n=no]
Figure 2: Illustration of various types of image operations
4
img
...Image Processing Fundamentals
2.2.2 Types of neighborhoods
Neighborhood operations play a key role in modern digital image processing. It is
therefore important to understand how images can be sampled and how that relates
to the various neighborhoods that can be used to process an image.
· Rectangular sampling ­ In most cases, images are sampled by laying a
rectangular grid over an image as illustrated in Figure 1. This results in the type of
sampling shown in Figure 3ab.
· Hexagonal sampling ­ An alternative sampling scheme is shown in Figure 3c and
is termed hexagonal sampling.
Both sampling schemes have been studied extensively [1] and both represent a
possible periodic tiling of the continuous image space. We will restrict our
attention, however, to only rectangular sampling as it remains, due to hardware and
software considerations, the method of choice.
Local operations produce an output pixel value b[m=mo,n=no] based upon the pixel
values in the neighborhood of a[m=mo,n=no]. Some of the most common
neighborhoods are the 4-connected  neighborhood  and  the  8-connected
neighborhood in the case of rectangular sampling and the 6-connected
neighborhood in the case of hexagonal sampling illustrated in Figure 3.
Figure 3a
Figure 3b
Figure 3c
Rectangular sampling
Rectangular sampling
Hexagonal sampling
4-connected
8-connected
6-connected
2.3 VIDEO P  ARAMETERS
We do not propose to describe the processing of dynamically changing images in
this introduction. It is appropriate--given that many static images are derived from
video cameras and frame grabbers-- to mention the standards that are associated
with the three standard video schemes that are currently in worldwide use ­ NTSC,
PAL, and SECAM. This information is summarized in Table 3.
5