Computer Vision Answers Key
Computer Vision Answers Key
• h(x, y): The degradation function. This represents things like blurring from motion or an out-of-
focus lens. Think of it as a "smudge" function that distorts the image.
• ∗: Convolution, which means the original image is combined with the degradation function.
• η(x, y): Additive noise—random variations caused by things like electronic sensor
imperfections, transmission interference, or atmospheric conditions.
In the frequency domain (which deals with the image in terms of its frequency components rather than
pixel intensities), this becomes:
Here, the capital letters represent the Fourier transforms of the respective spatial domain functions. This
frequency-domain representation is particularly useful because convolution in the spatial domain
becomes multiplication in the frequency domain, which is easier to handle computationally.
Key takeaway: The more accurately we know h(x, y) and η(x, y), the better we can estimate f(x, y). This
process is objective and based on modelling physical causes of degradation.
To convert a pixel from RGB to HSI, where R, G, B ∈ [0, 1], the following steps are used:
I = (R + G + B) / 3
Then,
If B ≤ G: H = θ
If B > G: H = 360° − θ
• S ∈ [0,1]
• I ∈ [0,1]
Note:
Edge detection is a fundamental image segmentation technique used to identify points in a digital image
where brightness changes sharply. These points typically represent object boundaries.
This method is based on detecting sharp changes in intensity by computing the gradient of the image.
The key ideas are:
a) Image Gradient:
The gradient of an image f(x, y) at a point gives the direction and rate of fastest increase in
intensity. It is a vector:
These values indicate how strong and in which direction the intensity changes, which helps detect
edges.
Summary:
Gradient-based edge detection detects intensity changes using the first derivative of the image. It’s
sensitive to noise, so smoothing and thresholding are commonly used. The gradient magnitude
shows edge strength, and its direction indicates orientation.
The Hit-or-Miss Transform (HMT) is a morphological operation used primarily for shape detection in
binary images. Unlike basic morphological operations like dilation or erosion, HMT utilizes two
structuring elements: one to probe the foreground (object) and one to probe the background.
Definition:
• Foreground: A
• Background: Ac (complement of A)
Let B1 be a structuring element for detecting a specific foreground pattern, and B2 for detecting a
background pattern.
• Detecting specific shapes or patterns (like corners, holes, or particular object outlines) in binary
images.
Example:
If you want to find a 3×3 cross shape within an image, you can define B1 to match the cross in the
foreground and B2 to match the surrounding background. The intersection of the two erosions pinpoints
where the pattern occurs.
Advantages:
• Can differentiate between similar structures based on foreground and background geometry.
Colour Generation Colours are created by adding light of Colours are created by subtracting light
Method R, G, B from white
Basis Used in devices that emit light (e.g., Used in devices that reflect light (e.g.,
monitors, cameras) printers, copiers)
Secondary Colours Cyan (G + B), Magenta (R + B), Red (M + Y), Green (C + Y), Blue (C + M)
Yellow (R + G)
White Representation (1, 1, 1) — full presence of light (0, 0, 0) — no absorption; full reflection of
R, G, B
Application Used in displays, scanners, digital Used in colour printing and inkjet devices
cameras
▪ Computer monitors
▪ Televisions
▪ Digital cameras
▪ Smartphones
o Used in colour reproduction systems that use pigments or dyes, such as:
▪ Inkjet printers
▪ Photocopiers
▪ Press printing (CMYK adds black (K) to compensate for imperfect colour inks)
o CMY is derived from RGB and is used for converting screen colours to printable output.