AI Topics Summary
AI Topics Summary
2. Information Extraction
Concept: Information extraction (IE) is the task of identifying and structuring specific pieces
of information from unstructured text data, such as names, dates, and relationships
between entities. Common applications include extracting addresses from websites or
weather conditions from reports.
Methods in Information Extraction:
Probabilistic Models:
Hidden Markov Models (HMM): Used for extracting sequences in noisy text data, such as
names or addresses. HMMs model a sequence of hidden states, allowing the system to
probabilistically “guess” the correct sequence based on observed data.
Conditional Random Fields (CRF): CRFs are advantageous for tasks that require contextual
awareness across multiple attributes, relaxing the independence assumptions in HMMs.
Example: In named entity recognition, HMMs might tag each word in a sentence as part of
an entity or not, while CRFs use the surrounding words to improve accuracy.
Edge Detection:
Concept: Detects points where image brightness changes significantly, corresponding to
potential object boundaries or depth changes.
Mathematics:
The brightness gradient across a cross-section perpendicular to an edge can be calculated
by differentiating the image intensity I(x). Locations where I'(x) (the derivative) is large
correspond to edges.
A Gaussian filter is applied to smooth the image, reducing noise. The Gaussian function is
defined as:
G(x) = 1/√(2πσ²) * exp(−x²/2σ²)
Applying the Gaussian filter at a pixel (x₀, y₀) smooths the image by computing a weighted
sum of neighboring pixel intensities:
I'(x₀, y₀) = Σ I(x, y) G(d)
Example: For an image of a desk with objects, edge detection identifies the object's outline
but may include noise edges. Later stages refine these detections.
Texture Analysis:
Concept: Texture is a visually repeating pattern, like stitches on fabric or spots on a leopard,
which can reveal surface characteristics. Some textures are periodic, while others, like
pebbles on a beach, follow only statistical regularity.
Application: Texture analysis is essential for detecting object boundaries in images with
intricate textures. For instance, the orientation histograms for a tiger’s fur and surrounding
grass differ, enabling the system to distinguish the animal from its background.
Optical Flow:
Concept: Optical flow describes the apparent motion of objects or the camera itself in a
video sequence. It measures pixel displacement over time, with motion recorded in terms of
pixels per second.
Mathematics:
To compute optical flow, a Sum of Squared Differences (SSD) method is often used. This
measures the similarity between pixel blocks centered at (x₀, y₀) at time t and candidate
pixels at (x₀ + Δx, y₀ + Δy) at time t + Δt:
SSD = Σ [I(x₀ + x, y₀ + y, t) - I(x₀ + x + Δx, y₀ + y + Δy, t + Δt)]²
Example: A moving race car in a video sequence shows high optical flow in the car and
slower flow in the background, useful for recognizing actions or estimating distances.
Image Segmentation:
Concept: Segmentation divides an image into regions of similar pixels, which helps in
identifying object boundaries by grouping regions based on brightness, color, or texture.
Example: Segmenting a landscape photo can separate regions of sky, water, and land.
Algorithms that create superpixels group pixels into homogeneous regions, simplifying
subsequent analysis.