U4
U4
Processing
• Offered to Final Year B.Tech. CSE
• By Dept. of C.Tech.
Unit 4 Topics
• Information Extraction and its approaches
• Information Retrieval
• Semantic Search
• Summarization
• Information Fusion
• Machine Translation
Information Extraction and its
approaches
Information Extraction (IE)
Information Extraction (IE) is an essential component of NLP for several reasons:
1. Structured Data from Unstructured Text: Much of the world's information is stored in
unstructured text form, such as articles, websites, documents, and social media posts. IE
enables the conversion of this unstructured text into structured data, making it usable
for various applications.
2. Data Organization and Summarization: IE techniques can extract key information from
large volumes of text, allowing for the organization and summarization of content. This is
valuable for tasks like creating document summaries, generating metadata, and
organizing data for analysis.
3. Automating Repetitive Tasks: Many NLP tasks involve manual and repetitive work, such
as cataloging information, tagging entities, or extracting facts from documents. IE
automates these tasks, saving time and reducing human error.
Information Extraction (IE)
Information Extraction (IE) is an essential component of NLP for several reasons:
4. Search and Retrieval: IE helps improve the accuracy and relevance of search engines and
recommendation systems. By extracting and indexing structured information from text, it
enhances the precision of search results and recommendations.
5. Entity Recognition and Linking: IE is used to identify and link entities (such as names of
people, places, organizations, etc.) in text. This is crucial for tasks like entity resolution,
disambiguation, and entity-based information retrieval.
Information Extraction (IE)
• Working with an enormous amount of text data is always hectic
and time-consuming.
• https://nanonets.com/blog/information-extraction/
Information Extraction
• What is information extraction?
– Information Extraction is the process of parsing through
unstructured data and extracting essential information into
more editable and structured data formats.
• Usually, we search for some required information when the data is digital or
manually check the same. But with information extraction NLP algorithms, we
can automate the data extraction of all required information such as tables,
company growth metrics, and other financial details from various kinds of
• After curating the data, we’ll then start applying the information
extraction NLP techniques, to process and build models around the
data.
Most common techniques for IE
• Tokenization
• Dependency Graphs
Tokenization
• Computers usually won't understand the language we speak or communicate
with. Hence, we break the language, basically the words and sentences, into
tokens and then load it into a program. The process of breaking down language
into tokens is called tokenization.
• Hence, with POS tagging we can use techniques that will provide the context of
words or tokens used to categorise them in specific ways.
Parts of Speech Tagging
• In parts of speech tagging, all the tokens in the text data get categorised into
different word categories, such as nouns, verbs, adjectives, prepositions,
determiners, etc.
import spacy
NLP = spacy.load("en_core_web_sm")
doc = NLP("Apple is looking at buying
U.K. startup for $1 billion")
for token in doc:
print(token.text, token.pos_)
Dependency Graphs
• Dependency graphs help us find relationships between neighbouring words using
directed graphs. This relation will provide details about the dependency type (e.g.
Subject, Object etc.).
– Hence, we’ll have to write different scripts to collect and store information in
one place.
– This is usually done by either using APIs on the web or building RPA
(Robotic Process Automation) pipelines.
Steps on how IE from text
2. Process Data
– After we collect the data, the next step is to process them.
– Usually, documents are two types: electronically generated (editable) and the other
non-electronically generated (scanned documents).
– For the electronically generated documents, we can directly send them into the
preprocessing pipelines.
– Still, we’ll need OCR to first read all the data from images and then send them into
preprocessing pipelines for the scanned copies.
– We can either use open-source tools like Tesseract or any online services like
Nanonets or Textract. After all the data is in editable or electronic format, we can then
apply to pre-process steps like Tokenization and POS tagging and then use data
loaders to load the data into the NLP information extraction models.
Steps on how IE from text
3. Choosing the right model
– As discussed in the above sections, choosing a suitable model mostly depends
on the type of data we’re working with. Today, there are several state-of-the-
art models we could rely on. Below are some of the frequently use open-
source models:
– Named Entity Recognition on CoNLL 2003 (English)
– Key Information Extraction From Documents: Evaluation And Generator
(GAN)
– Deep Reader: Information extraction from Document images via relation
extraction and Natural Language
– These are some of the information extraction models. However, these are
trained on a particular dataset. If we are utilising these on our models, we’ll
need to experiment on the hyperparameters and fine-tune the model
accordingly.
– The other way is to utilize the pre-trained models and fine-tuning them based
on our data. For Information Extraction from text, in particular, BERT models
are widely used.
Steps on how IE from text
4. Evaluation of the Model
• Evaluate the training process is crucial before we use the models in production.
This is usually done by creating a testing dataset and finding some key metrics:
• Accuracy: the ratio of correct predictions made against the size of the test data.
• Precision: the ratio of true positives and total predicted positives.
• Recall the ratio of true positives and total actual positives.
• F1-Score: harmonic mean of precision and recall.
• Different metrics take precedence when considering different use cases. In
invoice processing, we know that an increase in the numbers or missing an item
can lead to losses for the company. This means that besides needing a good
accuracy, we also need to make sure the false positives for money-related
fields are minimum, so aiming for a high precision value might be ideal. We also
need to ensure that details like invoice numbers and dates are always extracted
since they are needed for legal and compliance purposes. Maintaining a high
recall value for these fields might take precedence.
Steps on how IE from text
5. Deploying model in production
• The full potential of the NLP models only knows when they are deployed in
production.
• Today, as the world is entirely digital, these models are stored on cloud servers
with a suitable background. In most cases, Python is utilised as its more handy
programming language when it comes to Text data and machine learning. The
model is either exported as API or an SDK (software development kit) for
integrating with business tools.
• However, we need not build everything from scratch as there are several tools and
online services for this kind of use-cases. For example, Nanonets has a highly
accurate, fully trained invoice information extraction NLP model, and you can
directly integrate on our applications using APIs or supported SDKs.
Example - IE on ID card
https://nanonets.com/blog/information-extraction/#how-does-information-
extraction-work?
Information Retrieval
Information Retrieval (IR)
What is information retrieval?
• The system assists users in finding the information they require but it does not
explicitly return the answers of the questions.
• It informs the existence and location of documents that might consist of the
required information. The documents that satisfy user’s requirement are called
relevant documents. A perfect IR system will retrieve only relevant documents.
Basics of Information Retrieval (IR)
• It is clear from the above diagram that a user who needs information will have to
formulate a request in the form of query in natural language.
• Then the IR system will respond by retrieving the relevant output, in the form of
documents, about the required information.
Basics of Information Retrieval (IR)
• A user who needs information will have to formulate a request in the form
of a query in natural language. After that, the IR system will return output
by retrieving the relevant output, in the form of documents, about the
required information.
• Indexing
• Matching
Indexing
• Indexing is the process of preparing and organizing the collection of documents
or data to make it efficient for retrieval.
• This involves several tasks:
1. Tokenization:
• Breaking down text into individual words or tokens.
• For example, the sentence "I love programming" would be tokenized into
three tokens: "I," "love," and "programming."
2. Removing Frequent Words:
• Common words that don't contribute much to the meaning of the text,
such as "the," "and," "in," are often removed. These are known as stop
words.
3. Stemming:
• Reducing words to their root or base form. For example, "running" and
"ran" would both be stemmed to "run."
4. Creating an Inverted Index:
• This is a data structure that maps words or terms to the documents in
which they appear. It speeds up the retrieval process by allowing the
system to quickly locate documents containing specific terms.
Inverted Index
• An Inverted Index is a data structure used in information retrieval systems to
efficiently retrieve documents or web pages containing a specific term or set of
terms. In an inverted index, the index is organized by terms (words), and each
term points to a list of documents or web pages that contain that term.
• Example:
Inverted Index
2. Comparing with documents: The system then compares the processed query with the
indexed documents. It identifies documents that contain terms matching those in the
query.
3. Ranking: The retrieved documents are ranked by relevance. Various algorithms, such
as TF-IDF (Term Frequency-Inverse Document Frequency) or BM25, are used to
assign scores to documents based on how well they match the query.
4. Presenting results: Finally, the system presents the ranked list of documents to the
user, with the most relevant ones at the top.
Relevance
• Relevance refers to how effectively a document or collection
of documents that has been retrieved satisfies the user's
information needs.
• TF: It stands for Term Frequency which is simply the number of times a given
• IDF: It stands for Inverse Document Frequency which is a measure of the general
• F: The modeling framework for D, Q along with the relationship between them.
• F represents the modeling framework that combines the document and query
representations and captures the relationship between them. This modeling
framework could be a machine learning model, deep learning architecture, or any
other algorithm that processes the representations of documents and queries to
generate relevant rankings or similarity scores.
• R (q, di): A ranking or similarity function that orders the documents with respect to the
query.
Types of IR models
• Classic IR Model: This is described as the most basic and straightforward IR
model, grounded in mathematical information. The three traditional models
mentioned are Boolean, Vector, and Probabilistic.
– It's not a pre-arranged or pre-organized search but rather a flexible and tailored
retrieval process designed to meet the immediate needs of the user.
Classical problem in IR systems
• Ad-hoc Retrieval: Ad-hoc retrieval is described as a classical problem within the
information retrieval paradigm.
– It involves presenting a natural language query to retrieve relevant information. In
other words, it's the process of searching for information based on a user's query.
– Handling Irrelevant Information: After the user submits a query, the retrieved
information includes both relevant and non-relevant results. The non-relevant results,
in this context, are referred to as "ad hoc retrieval difficulties." These are documents
or pieces of information that don't satisfy the user's search criteria.
• Representation: In this step, the acquired data is organized and prepared for efficient
retrieval. It includes indexing, which involves creating data structures that facilitate quick
access to information. Representation techniques can include:
• Free-Text Terms: These are the actual words or phrases found in the documents.
• Controlled Vocabulary: This refers to a predefined set of terms used to categorize and
describe documents.
• Manual and Automatic Techniques: Both manual human input and automated algorithms
can be used to enhance the representation of data.
• As a response to the query, the set of documents that satisfied the Boolean expression are
retrieved.
– Q: It represents a Boolean expression, where terms are the index terms and operators are logical
products such as:
• AND,
– F: It represents a Boolean algebra over sets of terms as well as over sets of documents.
Boolean Model
• If we talk about the relevance feedback, then in the Boolean IR model the Relevance
prediction can be defined as follows:
• (( ˅ )˄ ˄˜ ℎ )
• User's Location: If the user is in New York City, context relevance might mean showing results
related to the Apple Store in Manhattan.
• User's Search History: If the user has been previously searching for tech-related topics, context
relevance might lead to results related to Apple Inc. and its products.
• User's Recent Behavior: If the user has recently clicked on articles about the latest iPhone
model, context relevance might prioritize results about Apple's latest product releases.
• In essence, context relevance considers the broader picture – the user's location,
preferences, and recent interactions – to ensure the search results are more meaningful in
the given context.
How does semantic search work?
• Intent Relevance: Intent relevance is all about understanding and aligning with the
specific purpose or intent behind the user's search. It ensures that the search results
directly address what the user is trying to achieve. Here's a more detailed explanation:
– Consider a user searching for "Apple." Intent relevance aims to grasp what the user wants to
do with that information. It could be:
• Purchase Intent: If the user is looking to buy an Apple product, intent relevance would
prioritize showing links to online or local stores where Apple products are available.
• Research Intent: If the user wants to learn more about Apple's history and innovations, intent
relevance would show articles, Wikipedia pages, and informative resources about Apple Inc.
• Technical Support Intent: If the user needs help with an Apple device or service, intent
relevance might provide links to Apple's official support pages or forums.
• In this way, intent relevance ensures that the search results align with the specific goal or
intention of the user's search. It helps deliver results that best serve the user's needs and
objectives.
How does semantic search work?
Semantic search vs. keyword search
• The difference between semantic search and keyword search is that keyword search
returns results that match words to words, words to synonyms, or words to similar words.
• Semantic search looks to match the meaning of the words in the query. In some cases,
semantic search might not generate results with direct word matches, but it will match the
user’s intent.
• Keyword search engines use query expansion or relaxation tools like synonyms or word
omission. They also use natural language processing and understanding tools like typo
tolerance, tokenization, and normalization. On the other hand, semantic search is able to
return query results that match meaning through the use of vector search.
• Consider "chocolate milk." A semantic search engine will distinguish between “chocolate
milk” and “milk chocolate.” Though the keywords in the query are the same, the order in
which they are written affects the meaning. As humans, we understand that milk chocolate
refers to a variety of chocolate, whereas chocolate milk is chocolate-flavored milk.
Semantic search vs. keyword search
• Semantic search offers results based on the user’s geographical context, the user’s past
search history, and user intent.
• With a better understanding of user intent, semantic search can respond to a query like
"Creuset vs. Staub dutch ovens" with content that prioritizes product comparisons because
that is the user’s intent. Semantic search will recognize the intent behind “best Staub
deals” or "Creuset discounts" as intent to purchase and offer responses accordingly.
• Another example is predictive text. As you type a query into a search bar, it uses semantic
search to complete your query and suggest relevant search terms based on context,
common searches, and past search history.
Benefits of semantic search
• Easier to use for customers
Customers might not remember jargon, or recall specific product
names. Semantic search enables customers to input vague search
queries and get specific results. Customers can also search using a
description to discover its name. For example, you can discover a
song by searching for the lyrics that you know and find the title.
– Extractive Summarization
– Abstractive Summarization:
Approaches in auto summarization
• Extraction Summarization:
This approach entails the method to extract keywords and phrases
from sentences and then joining them to produce a compact
meaningful summary.
• Abstractive Summarization:
In this summary generator, algorithms are developed in such a way to
reproduce a long text into a shorter one by NLP. It retains its
meaning but changes the structure of sentences.
Extractive Vs Abstractive Summarization
ASPECT EXTRACTIVE SUMMARIZATION ABSTRACTIVE
SUMMARIZATION
Definition Selects and compiles important Generates a summary using
sentences or phrases from the original natural language generation
text techniques
Content Reproduces content from the original May include original content not
text present in the original text
Output Sentence or phrase-based More fluent, human-like language
store the text data and its frequency in a dictionary. After that, we tokenize our
text data. The sentences which contain more high frequency words will be kept
word in the source text. It may use a simple dictionary or data structure to
keep track of word occurrences. This step can be implemented with a basic
algorithm that iterates through the text and updates the word frequencies.
Algorithms used for Extractive Summarization
documents as vectors. This is often done using techniques like TF-IDF (Term
2. Vector Normalization: Normalize the vectors. This means dividing each vector
by its magnitude (also called the vector's length). This step ensures that the
length of the vectors does not affect the cosine similarity calculation.
• ||A|| and ||B|| represent the magnitudes (or lengths) of vectors A and B,
respectively. These magnitudes are the square roots of the sum of the
squares of the individual vector components.
1. Sequence-to-Sequence models
2. Pointer-Generator Networks
3. Transformer-based models
Sequence-to-Sequence models
• A Seq2Seq model is a type of neural network architecture used NLP and other sequence-
related tasks. It is designed to handle sequences of data, where the length of the input and
output sequences can vary. Sequence-to-sequence models, often implemented using RNNs
or more modern variants like LSTMs and GRUs, utilize an encoder-decoder architecture.
1. Encoder (process inputs): The encoder component takes an input sequence and
processes it step by step, usually from left to right. It transforms the input sequence
into a fixed-length vector called the "context" or "thought" vector. This context vector
represents the entire input sequence in a compressed form and captures the important
information from the input.
2. Decoder (generate output): The decoder component takes the context vector produced
by the encoder and generates an output sequence, often one step at a time. The decoder
can be autoregressive, meaning it generates each output token while conditioning on
the previously generated tokens.
Sequence-to-Sequence models
• Seq2Seq models are particularly versatile and have been used for a wide range of
recognition.
• Machine Translation: Seq2Seq models are known for their effectiveness in translating
text from one language to another. The encoder processes the input text in the source
language, and the decoder generates the translated text in the target language.
• Text Summarization: In abstractive text summarization, the encoder processes the input
• Speech Recognition: Seq2Seq models can convert spoken language (audio) into text.
The encoder processes the audio waveform, and the decoder generates a corresponding
https://towardsdatascience.com/understanding-encoder-decoder-sequence-to-
sequence-model-679e04af4346
https://www.cloudskillsboost.google/course_sessions/5521175/video/383848
Pointer-Generator Networks
• Pointer-Pointer-Generator Networks are a neural architecture, built on RNNs or
transformers, that processes input and generates output.
5. Word Generation: In addition to copying words from the source text, the decoder
generates new words and phrases when necessary to create a coherent summary. It does so
by learning how to predict which words to generate based on the context.
• Unlike traditional recurrent neural networks (RNNs) and convolutional neural networks
(CNNs), transformers rely on self-attention mechanisms to process sequences of data.
• These models have shown promising results for abstractive summarization tasks, such as
BERT, GPT-2, T5, etc.
• Self-Attention Mechanism: The self-attention mechanism allows the model to consider the
relationships and dependencies between all words or tokens in a sequence
simultaneously. This enables transformers to capture complex contextual information,
including long-range dependencies, which is crucial for understanding the meaning of text.
• GPT-2 (Generative Pre-trained Transformer 2): GPT-2 is a generative model that uses
transformers to generate human-like text. It's capable of abstractive summarization by
generating coherent and contextually relevant summaries. GPT-2 has been applied to a
wide range of NLP tasks, including summarization.
• T5 (Text-to-Text Transfer Transformer): T5 is a model that frames all NLP tasks as text-
to-text tasks. It has shown remarkable results in various tasks, including summarization. By
treating summarization as a text-to-text task, T5 can generate abstractive summaries
effectively.
Sequence-to-Sequence vs Pointer-Generator vs Transformer-
based models
Sequence-to-Sequence vs Pointer-Generator vs Transformer-
based models
Information Fusion
Information Fusion
• Information fusion in NLP is the process of combining
information from multiple sources to create a more
comprehensive and accurate understanding of a situation
NLP tasks.
• These sources can include text, speech, images, video, and various other types of data.
Example
• A machine translation system can use information fusion to
combine the translations of multiple different models to
produce a more accurate translation.
For example, in a multimedia search task, you might want to combine text descriptions
2. Redundancy Reduction: By integrating information from multiple sources, you can reduce
redundancy and improve the reliability of the extracted information. For instance, you
might use both text and speech data to transcribe spoken language, which can help in
reducing errors.
3. Context Enhancement: Information fusion can provide additional context to a given piece
of information. For instance, combining information from different news sources can give
social media posts or user reviews to get a more accurate understanding of public sentiment
texts can lead to more accurate translations, especially when dealing with languages with
• Early Fusion: Combining data from different sources at the very beginning of the
processing pipeline. For example, fusing text and images as input for a multimodal model.
• Late Fusion: Processing data from different sources independently and then combining the
results. For instance, analyzing text and speech separately and then integrating the findings.
• Decision-Level Fusion: Combining the outputs or decisions from multiple models, each
trained on different data sources.
• Contextual Fusion: Incorporating contextual information from one source to aid in the
processing of another source.
Information Fusion
• In summary, Information fusion in NLP involves the integration of data from various
and extract more valuable insights from the data. It plays a crucial role in addressing
• https://www.v7labs.com/blog/multimodal-deep-learning-guide
Single and Multi-document
Summarization – Question
Answering
Single and Multi-document
Summarization
Single and Multi-document Summarization
• Single-Document Summarization:
2. Content Abstraction: It generates a concise summary of the source text, often in a few
sentences or paragraphs, while retaining the core meaning.
1. Aggregating Information: It aims to provide a comprehensive summary that captures the most
important details from multiple source documents, ensuring that the main points from all documents
are included.
2. Redundancy Reduction: This technique addresses the issue of redundancy, ensuring that repetitive
information found in multiple documents is summarized efficiently.
3. Content Fusion: Information from different sources is often merged to create a coherent and
informative summary that provides a more comprehensive view of the topic.
or a set of documents. When a question is asked, the system matches it with similar
questions or extracts relevant information from the database to provide an answer.
• Example: A frequently asked questions (FAQ) chatbot that matches user questions to
predefined answers based on keywords or similarity.
Generative-based QA:
Generative-based QA models generate answers from scratch based on the
Retrieval-based QA:
System: "The capital of France is Paris."
Generative-based QA:
System: "The capital of France is Paris."
QA systems can be applied in various domains, including customer support,
information retrieval, educational platforms, and more. They can range from simple
rule-based systems to advanced machine learning models like transformers.
• To have a conversation with your AI, need a few pre-trained tools which
can help you build an AI chatbot system.
interface.
• Traditional chatbots: They are driven by system and automation, mainly through
scripts with minimal functionality and the ability to maintain only system context.
system and humans. They have the ability to maintain both system and task contexts.
• Future chatbot: They can communicate at multiple levels with automation at the
system level. They have the ability to maintain the system, task, and people contexts.
• The bots can handle simple queries but fail to manage complex
ones.
• Self-learning bots are the ones that use some Machine Learning-
based approaches and are definitely more efficient than rule-based
bots.
set of predefined answers or a knowledge base. They select the most appropriate
response based on the input query. This approach is useful for fact-based or
frequently asked questions where answers are readily available in the training data
or knowledge base.
on the fly. They use machine learning models, often based on deep learning
techniques, to generate text that is contextually relevant to the input query. This
answers and is particularly useful for tasks like chatbots and content generation.
Applications of Chatbots
• Virtual reception assistant
• Virtual operations assistant [example: Jarvis from the movie Iron Maiden]
• Chatbot asks the user to type in the chat window using the NLTK
converse function.
• Bot understands what the user has typed in the chat utility
window using NLTK chat pairs and reflections function.
Data pre-processing
Text case [upper or lower] handling
• The system usually trained in such a way that, if a certain input had
given by the user, the response should be a particular one.
• However this rule based system are not able to produce output
when a query which are not defined in the rule set is asked.
AI BASED SYSTEM
• Artificial Intelligence Based System generally uses the concept
of machine learning approaches to produce responses for the
input which is given by the user.
• While the other two methods produces output on the basis of the rule
defined or by considering matching from database, generative model needs
large amount of data to get trained and from that the system try to generate
new responses for a new query.
• The most basic form of NLU is parsing, which takes text written
in natural language and converts it into a structured format that
computers can understand.
Lexical Analysis: Next, the tokens are placed into a dictionary that includes
their part of speech (for example, whether they're nouns or verbs). It also
includes identifying phrases that should be placed in a separate database for
later use.
Syntactic Analysis: The tokens are analyzed for their grammatical structure.
It includes identifying each word's roles and whether there's any ambiguity
between multiple interpretations of those roles.
How Does Natural Language Understanding
Work?
Natural Language Understanding(NLU) Example
1. Alice is swimming against the current.
• In the first sentence, the word, current is a noun. The verb that precedes
it, swimming, provides additional context to the reader, allowing us to
conclude that we are referring to the flow of water in the ocean.
• The second sentence uses the word current, but as an adjective. The
noun it describes, version, denotes multiple iterations of a report,
enabling us to determine that we are referring to the most up-to-date
status of a file.
Natural Language Understanding Applications
• Data Capture
• Customer Support
• Chatbots
• Virtual assistants
Natural Language Generation
(NLG)
Natural Language Generation (NLG)
• Data understanding.
• Sentence aggregation.
• Grammatical structuring.
• Language presentation.
Natural Language Generation Applications
• One of the most notable advantages of machine translation is its ability to rapidly
translate large volumes of text.
• However, this technology has come a long way since the mid-20th century.
• It works by alluding to statistical models that depend on the investigation of huge volumes
of bilingual content. It expects to decide the correspondence between a word from the
source language and a word from the objective language. A genuine illustration of this is
Google Translate.
• Presently, SMT is extraordinary for basic translation, however its most noteworthy
disadvantage is that it doesn't factor in context, which implies translation can regularly be
wrong or you can say, don't expect great quality translation. There are several types of
statistical-based machine translation models which are: Hierarchical phrase-based
translation, Syntax-based translation, Phrase-based translation, Word-based translation.
• https://www.analyticssteps.com/blogs/4-types-machine-translation-nlp
4 Types of Machine Translation
2. Rule based MT
• HMT, as the term demonstrates, is a mix of RBMT and SMT. It uses a translation
memory, making it unquestionably more successful regarding quality.
Nevertheless, even HMT has a lot of downsides, the biggest of which is the
requirement for enormous editing, and human translators will also be needed.
There are several approaches to HMT like multi-engine, statistical rule
generation, multi-pass, and confidence-based.
4 Types of Machine Translation
4. Neural Machine Translation or NMT
• NMT is a type of machine translation that relies upon neural network models
(based on the human brain) to build statistical models with the end goal of
translation. The essential advantage of NMT is that it gives a solitary system that
can be prepared to unravel the source and target text. Subsequently, it doesn't rely
upon specific systems that are regular to other machine translation systems,
particularly SMT.