0% found this document useful (0 votes)
13 views

All Units Notes - IM

Uploaded by

ssubitha2003
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
13 views

All Units Notes - IM

Uploaded by

ssubitha2003
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 185

BA 4106 Information Management

BA4106 INFORMATION MANAGEMENT L T P C 3 0 0 3

COURSE OBJECTIVES:
To understand the importance of information in business
To know about the recent information systems and technologies.

UNIT I INTRODUCTION 9

Data, Information, Information System, evolution, types based on functions and hierarchy,
Enterprise and functional information systems.

UNIT II SYSTEM ANALYSIS AND DESIGN 10

System development methodologies, Systems Analysis and Design, Data flow Diagram (DFD),
Decision table, Entity Relationship (ER), Object Oriented Analysis and Design (OOAD), UML
diagram.

UNIT III DATABASE MANAGEMENT SYSTEMS 8

DBMS – types and evolution, RDBMS, OODBMS, RODBMS, Data warehousing, Data Mart,
Data mining.

UNIT IV INTEGRATED SYSTEMS, SECURITY AND CONTROL 9

Knowledge based decision support systems, Integrating social media and mobile technologies in
Information system, Security, IS Vulnerability, Disaster Management, Computer Crimes,
Securing the Web.

UNIT V NEW IT INITIATIVES 9

Introduction to Deep learning, Big data, Pervasive Computing, Cloud computing, Advancements
in AI, IoT, Block chain, Crypto currency, Quantum computing
Total: 45 Periods

REFERENCES:

1. Robert Schultheis and Mary Sumner, Management Information Systems – The Manager’s
View, Tata McGraw Hill, 2008.
2. Kenneth C. Laudon and Jane P Laudon, Management Information Systems – Managing the
Digital Firm, 15 th edition, 2018.
3. Panneerselvam. R, Database Management Systems, 3rd Edition, PHI Learning, 2018.
Unit I: Introduction
Syllabus: Data, Information, Information System, evolution, types based on functions and
hierarchy, Enterprise and functional information systems.
Data Vs Information
Data can be described as unprocessed facts and figures. Plain collected data as raw facts cannot
help in decision-making. However, data is the raw material that is organized, structured, and
interpreted to create useful information.
Data is defined as 'groups of non-random symbols in the form of text, images, voice
representing quantities, action and objects'. Data are only the raw facts, the material for
obtaining information.
Information is interpreted data; created from organized, structured, and processed data in a
particular context.
Knowledge is the human expertise stored in a person’s mind, gained through experience, and
interaction with the person’s environment. So information when combined with (a manager’s)
insight, experience and expertise, becomes knowledge with which stronger decisions can be
made.
Example: The number 36 is data. Knowing that 36 is my age is information. Information that
the data represents my age is essential knowledge that is key in an information system.
Intelligence It requires ability to sense the environment, to make decisions, and to control
action. It is a step ahead and uses information and knowledge to drive business decisions.
Intelligence is decision-support. It’s a tool for making predictions about the future in order to
take a course of action that improves outcomes.
Systems a system is a set of components (subsystems) that operate together to achieve certain
objectives.
Types of System
Physical or Abstract: Physical system is tangible entities that may be static or dynamic
in nature. Abstract system is conceptual or non-physical. The abstract is
conceptualization of physical situations.
Open and Closed: An open system continually interacts with its environment. It
receives input from the outside and delivers output to outside. A closed system is isolated
from environment influences.
Sub System and Super System: Each system is part of a large system. The business
firm is viewed as the system or total system when focus is on production, distribution of
goal and sources of profit and income. The total system consists of all the objects,
attributes and relationship necessary to accomplish an objective given a number of
constraints. Sub systems are the smaller systems within a system. Super system denotes
extremely large and complex system
Permanent and Temporary System: A permanent system is a system enduring for a
time span that is long relative to the operation of human. Temporary system is one
having a short time span.
Natural and Man Made System: System which is made by man is called man made
system. Systems which are in the environment made by nature are called natural system.
Deterministic and Probabilistic: A Deterministic system is one in which the occurrence
of all events is perfectly predictable. If we get the description of the system state at a
particular time, the next state can be easily predicted. Probabilistic system is one in
which the occurrence of events cannot be perfectly predicted.
Information System (IS)
An information system is the group of procedures and resources used to gather, store, process
and communicate the information needed in an organization.
Information systems use data stored in computer databases to provide needed information.
1. Information systems capture data from the organization (internal data) and its
environment (external data). They converts the data into a meaningful information.
2. They store the database items over an extensive period of time.
3. When specific information is needed, the appropriate data items are manipulated as
necessary, and the user receives the resulting information.
Computer Based Information System (CBIS) - It is the subset of the information system that
automates information management using computers.
Information Technology
Information technology falls under the IS umbrella but deals with the technology involved in the
systems themselves. Information technology can be defined as the study, design,
implementation, support or management of computer-based information systems.
IT typically includes hardware, software, databases and networks. Information technology often
governs the acquisition, processing, storage and dissemination of digitized information, or data,
generated through the disciplines of computing and telecommunications. Information technology
focuses on managing technology and improving its utilization to advance the overall business
goals.

Information Systems vs. Information Technology


The difference between information systems and information techno logy is that Information
system incorporates the technology, people and processes involved with information.
Information technology is the design and implementation of information, or data, within the
information system.
Information Systems is about the creation, sharing, and distribution of information that is
generated by computers and their users. While people who work in this area do work with
computers and software and need access to networks, they are primarily concerned with
extracting data from various sources in order to accomplish various tasks.
Information technology is more about the hardware like computers, monitors, mice, keyboards,
printers as well as the internal components that make up these devices, software applications and
telecommunications. Very often, people who work in information technology work behind the
scenes to support those who use those devices, or are those who create the software, web pages,
and build networks.
While both information technology and information systems will involve working with others,
IS professionals are typically more integrated in using technology and other systems to
accomplish business objectives. IT professionals, while serving an important function to the
overall organization, are more focused on the machines and hardware and software systems.
Components of Information System
Information systems consist of the following general components:
1. Hardware 4. People
2. Software 5. Communication Networks
3. Databases 6. Procedures

Hardware: It refers to physical equipment used for input, output and processing. Hardware
refers to the computers themselves, along with any and all peripherals, including servers,
routers, monitors, printers and storage devices.

Software: The programs/ application program used to control and coordinate the hardware
components. It is used for analysing and processing of the data. These programs include a set
of instruction used for processing information .

(i) Systems (Operating) Software (ii) Applications Software.

 Systems (Operating) Software: An operating system is the principal system software. It


manages all the hardware of a computer system and provides an interface through which
the system's user can deploy computer hardware.
 Application Software: Are programs that directly assist end users in doing their work.
They are purchased as ready-to-use packages. Applications software directly assists end
users in doing their work.
Database is an organized collection of interrelated data reflecting a major aspect of a firm's
activities.
Procedures "procedures are to people like software is to hardware." procedures are the rules,
descriptions and instructions for how things are done. In computer-based information systems,
procedures are frequently covered in instruction or user manuals that describe how to use the
hardware, software and data.
Network: Networks resources refer to the telecommunication networks like the intranet,
extranet and the internet. These resources facilitate the flow of information in an information
system. Networks consists of both the physicals devises such as networks cards, routers, hubs
and cables. Networks include communication media, and Network Support.
People It is people who design and operate the software, input the data, build the hardware and
keep it running, write the procedures and it is ultimately people who determine the success or
failure of a CBIS. People are the most often overlooked and most important part of a computer-
based information system.

According to Davis and Olson −


"Information is a data that has been processed into a form that is meaningful to recipient and is
of real or perceived value in the current or the prospective action or decision of recipient."

Information, Knowledge and Business Intelligence


Professor Ray R. Larson of the School of Information at the University of California, Berkeley,
provides an Information Hierarchy, which is −
 Data − The raw material of information.
 Information − Data organized and presented by someone.
 Knowledge − Information read, heard, or seen, and understood.
 Wisdom − Distilled and integrated knowledge and understanding.
Information/Data Collection Techniques
The most popular data collection techniques include −
 Surveys − A questionnaires is prepared to collect the data from the field.
 Secondary data sources or archival data: Data is collected through old records,
magazines, company website etc.
 Objective measures or tests − An experimental test is conducted on the subject and the
data is collected.
 Interviews − Data is collected by the system analyst by following a rigid procedure and
collecting the answers to a set of pre-conceived questions through personal interviews.
Classification of Information
Information can be classified in a number of ways and in this chapter, you will learn two of the
most important ways to classify information.

Classification by Characteristics
 Action vs. non action information: Action information is active information that causes
an activity or operation, while the information that communicates only when the status is
applied without any operation is called non-action.
 Recurring vs. non recurring information: The information that is generated in regular
intervals is called recurring information, whereas non-repetitive in nature is called non-
recurring information.
 Internal vs. external information: all information that produced from internal sources
of any organization is called internal information, though all information that produced
from external sources of any organization is called external information.
Classification by Hierarchy
Information used in business for decision-making is generally categorized into three types −

 Strategic Information (Top level) − Strategic information is concerned with long term
policy decisions that defines the objectives of a business and checks how well these
objectives are met. For example, acquiring a new plant, a new product, diversification of
business etc, comes under strategic information.

 Tactical Information (Middle level) − Tactical information is concerned with the


information needed for exercising control over business resources, like budgeting,
quality control, service level, inventory level, productivity level etc.

 Operational Information (Low level) − Operational information is concerned with


plant/business level information and is used to ensure proper conduction of specific
operational tasks as planned/intended. Various operator specific, machine specific and
shift specific jobs for quality control checks comes under this category.

Classification by Application
In terms of applications, information can be categorized as −

 Planning Information − These are the information needed for establishing standard
norms and specifications in an organization. This information is used in strategic,
tactical, and operation planning of any activity. Examples of such information are time
standards, design standards.

 Control Information − This information is needed for establishing control over all
business activities through feedback mechanism. This information is used for controlling
attainment, nature and utilization of important processes in a system. When such
information reflects a deviation from the established standards, the system should induce
a decision or an action leading to control.

 Knowledge Information − Knowledge is defined as "information about information".


Knowledge information is acquired through experience and learning, and collected from
archival data and research studies.

Quality of Information
Information is a vital resource for the success of any organization. Future of an organization lies
in using and disseminating information wisely. Good quality information placed in right context
in right time tells us about opportunities and problems well in advance.
Good quality information − Quality is a value that would vary according to the users and uses of
the information. Let us generate a list of the most essential characteristic features for
information quality −
 Reliability − It should be verifiable and dependable.
 Timely − It must be current and it must reach the users well in time, so that important
decisions can be made in time.
 Relevant − It should be current and valid information and it should reduce uncertainties.
 Accurate − It should be free of errors and mistakes, true, and not deceptive.
 Sufficient − It should be adequate in quantity, so that decisions can be made on its basis.
 Unambiguous − It should be expressed in clear terms. In other words, in should be
comprehensive.
 Complete − It should meet all the needs in the current context.
 Unbiased − It should be impartial, free from any bias. In other words, it should have
integrity.
 Explicit − It should not need any further explanation.
 Comparable − It should be of uniform collection, analysis, content, and format.
 Reproducible − It could be used by documented methods on the same data set to
achieve a consistent result.
Need & Objective of Information
Information processing beyond doubt is the dominant industry of the present century. Following
factors states few common factors that reflect on the needs and objectives of the information
processing −
 Increasing impact of information processing for organizational decision making.

 Dependency of services sector including banking, financial organization, health care,


entertainment, tourism and travel, education and numerous others on information.
 Changing employment scene world over, shifting base from manual agricultural to
machine-based manufacturing and other industry related jobs.
 Information revolution and the overall development scenario.
 Growth of IT industry and its strategic importance.
 Strong growth of information services fuelled by increasing competition and reduced
product life cycle.
 Need for sustainable development and quality life.
 Improvement in communication and transportation brought in by use of information
processing.
 Use of information processing in reduction of energy consumption, reduction in
pollution and a better ecological balance in future.
 Use of information processing in land record managements, legal delivery system,
educational institutions, natural resource planning, customer relation management and
so on.
In a nutshell −
 Information is needed to survive in the modern competitive world.
 Information is needed to create strong information systems and keep these systems up to
date.
Implications of Information in Business
Information processing has transformed our society in numerous ways. From a business
perspective, there has been a huge shift towards increasingly automated business processes and
communication. Access to information and capability of information processing has helped in
achieving greater efficiency in accounting and other business processes.

A complete business information system, accomplishes the following functionalities −


 Collection and storage of data.
 Transform these data into business information useful for decision making.
 Provide controls to safeguard data.
 Automate and streamline reporting.
The following list summarizes the five main uses of information in businesses organizations −
 Planning − At the planning stage, information is the most important ingredient in
decision making. Information at planning stage includes that of business resources,
assets, liabilities, plants and machineries, properties, suppliers, customers, competitors,
market and market dynamics, fiscal policy changes of the Government, emerging
technologies, etc.
 Recording − Business processing these days involves recording information about each
transaction or event. This information collected, stored and updated regularly at the
operational level.
 Controlling − A business need to set up an information filter, so that only filtered data is
presented to the middle and top management. This ensures efficiency at the operational
level and effectiveness at the tactical and strategic level.
 Measuring − A business measures its performance metrics by collecting and analyzing
sales data, cost of manufacturing, and profit earned.
 Decision-making − MIS is primarily concerned with managerial decision-making,
theory of organizational behavior, and underlying human behavior in organizational
context. Decision-making information includes the socio-economic impact of
competition, globalization, democratization, and the effects of all these factors on an
organizational structure.
Evolution of Information System
An information system is a combination of processes, hardware, trained personnel, software,
infrastructure and standards that are designed to create, modify, store, manage and distribute
information to suggest new business strategies and new products. It leads to efficient work
practices and effective communication to make better decisions in an organization. There has
been a significant evolution of Information System function over the past few decades.

The evolution of Information System function can be summarized as follows:


1950 – 1960 1960 – 1970 1970 – 1980 1980 – 1990 1990 – 2000 2000 – Present

Management Decision Executive Knowledge


Data Processing E-Business
Reporting Support Support Management

Pre-specified
Provide both Supports the Greater
Collects, stores, reports and
Interactive ad-hoc internal and creation, connectivity,
modifies and displays to
support for the external organization and higher level of
retrieve day-to- support
decision-making information dissemination of integration
day transactions business
process relevant to the business across
of an decision-
strategic goals of knowledge applications
organization making
the organization

Helps senior Help available Helps global e-


Help workers Helps middle
managers Helps Executives enterprise wide business
managers

Evolution of Information System

1950 – 1960: Electronic Data Processing, Transaction Processing System

During this period, the role of IS was mostly to perform activities like transaction
processing, recordkeeping and accounting. IS was mainly used for Electronic Data Processing
(EDP). EDP is described as the use of computers in recording, classifying, manipulating, and
summarizing data. It is also called information processing or automatic data processing.

Transaction Processing System (TPS) was the first computerized system developed to process
business data. TPS was mainly aimed at clerical staff of an organisation. The early TPS used
batch processing data which was accumulated over a period and all transactions were processed
afterward. TPS collects, stores, modifies and retrieves day-to-day transactions of an organization.
Usually, TPS computerize or automate an existing manual process to allow for faster processing,
improved customer service and reduced clerical costs.

Examples of outputs from TPS are cash deposits, automatic teller machine (ATM), payment
order and accounting systems. TPS is also known as transaction processing or real-time
processing.
1960 to 1970: Management Information Systems

During this era, the role of IS evolved from TPS to Management Information Systems (MIS).
MIS process data into useful informative reports and provide managers with the tools to organize
evaluate and efficiently manage departments within an organization. MIS delivers information in
the form of displays and pre-specified reports to support business decision-making. Examples of
output from MIS are cost trend, sales analysis and production performance reporting systems.
Usually, MIS generates three basic types of information which are:

 Detailed information reports typically confirm transaction-processing activities. A


detailed Order Report is an example of a detailed report.
 Summary information establishes data into a format that an individual can review quickly
and easily.
 Exception information report information by filtering data that is an exception inventory
report. Exception reports help managers save time because they do not have to search
through a detailed report for exceptions.

This period also marked the development when the focus of organizations shifted slowly from
merely automating basic business processes to consolidating the control within the data
processing function.

1970 to 1980: Decision Support Systems


In this era, a major advancement was an introduction of the personal computers (PC). With the
introduction of PCs, there was the distribution of computing or processing power across the
organization. IS function associated strongly with management rather than a technical approach
in an organisation. The role focused on “interactive computer-based system” to aid decision-
makers in solving problems.
This new role of information systems to provide interactive ad-hoc support for the decision-
making process to managers and other business professionals is called Decision Support Systems
(DSS). DSS serve the planning, management and operations level of an organization usually
senior management.
DSS uses data from both internal and/or external sources. Internal sources of data might include
inventory, sales, manufacturing or financial data from an organization’s database. External
sources could include pricing, interest rates, population or trends. Managers use DSS to
manipulate the data to help with decisions. Examples of DSS are projected revenue figures based
on new product sales assumptions, product pricing and risk analysis systems.

1980 to 1990: Executive Information Systems

This period gave rise to departmental computing due to many organisations purchasing their own
hardware and software to suit their departmental needs. Instead of waiting for indirect support of
centralized corporate service department, employees could use their own resources to support
their job requirements. This trend led to new challenges of data incompatibility, integrity and
connectivity across different departments. Further, top executives were neither using DSS nor
MIS hence executive information systems (EIS) or executive support systems (ESS) were
developed.

EIS offers decision making facilities to executives through providing both internal and external
information relevant to meeting the strategic goals of the organization. These are sometimes
considered as a specific form of DSS. Examples of the EIS are systems for easy access to actions
of all competitors, economic developments to support strategic planning and analysis of business
performance.

1990 to 2000: Knowledge Management Systems


During this era, the rapid growth of the intranets, extranets, internet and other interconnected
global networks dramatically changed the capabilities of IS in business. It became possible to
circulate knowledge to different parts of the world irrespective of time and space.
This period also saw an emergence of enterprise resource planning (ERP) systems. ERP is an
organization-specific form of a strategic information system that incorporates all components of
an organisation including manufacturing, sales, resource management, human resource planning
and marketing.
Moreover, there was a breakthrough in the development and application of artificial intelligence
(AI) techniques to business information systems. Expert systems (ES) and knowledge
management systems (KMS) interconnected to each other.
Expert systems (ES) are a computer system that mimics the decision-making ability of human
experts. For example, systems making financial forecasts, diagnosing human illnesses and
scheduling routes for delivery vehicles.
Knowledge management system (KMS) is an IT system that stores and retrieves knowledge to
support creation, organization and dissemination of business knowledge within the enterprise.
Examples of KMS are feedback database and helpdesk systems.
ES uses data from Knowledge Management Systems to generate desirable information system’s
output for example loan application approval system.
2000 – present: E-Business

The Internet and related technologies and applications changed the way businesses operate and
people work. Information systems functions in this period are still the same just like 50 years ago
doing records keeping, reporting management, transactions processing, support management and
managing processes of the organization. It is used to support business process, decision making
and competitive advantage.

The difference is greater connectivity across similar and dissimilar system components. There is
great network infrastructure, higher level of integration of functions across applications and
powerful machines with higher storage capacity. Many businesses use Internet technologies and
web-enable business processes to create innovative e-business applications. E-business is simply
conducting business process using the internet.
Information Systems based on Functions & Hierarchy
As most organizations are hierarchical, the way in which the different classes of information
systems are categorized tends to follow the hierarchy. This is often described as "the pyramid
model" because the way in which the systems are arranged mirrors the nature of the tasks found
at various different levels in the organization.

For example,

Three level pyramid model based on the type of decisions taken at different levels in the organization

What are the most common types of information system in an organization?

Basing the classification on the people who use the information system means that many of the
other characteristics such as the nature of the task and informational requirements, are taken into
account more or less automatically.

Four level pyramid model based on the different levels of hierarchy in the organization
Transaction Processing System (TPS)
Transaction processing systems are used to record day to day business transactions of the
organization. They are used by users at the operational management level. The main objective of
a transaction processing system is to answer routine questions such as;
 How printers were sold today?
 How much inventory do we have at hand?
 What is the outstanding due for John Doe?
By recording the day to day business transactions, TPS system provides answers to the above
questions in a timely manner.
 The decisions made by operational managers are routine and highly structured.
 The information produced from the transaction processing system is very detailed.
For example, banks that give out loans require that the company that a person works for should
have a memorandum of understanding (MoU) with the bank. If a person whose employer has a
MoU with the bank applies for a loan, all that the operational staff has to do is verify the
submitted documents. If they meet the requirements, then the loan application documents are
processed. If they do not meet the requirements, then the client is advised to see tactical
management staff to see the possibility of signing a MoU.
Examples of transaction processing systems include;
 Point of Sale Systems – records daily sales
 Payroll systems – processing employees salary, loans management, etc.
 Stock Control systems – keeping track of inventory levels
 Airline booking systems – flights booking management
Management Information System (MIS)
Management Information Systems (MIS) are used by tactical managers to monitor the
organization's current performance status. The output from a transaction processing system is
used as input to a management information system.
The MIS system analyzes the input with routine algorithms i.e. aggregate, compare and
summarizes the results to produced reports that tactical managers use to monitor, control and
predict future performance.
For example, input from a point of sale system can be used to analyze trends of products that are
performing well and those that are not performing well. This information can be used to make
future inventory orders i.e. increasing orders for well-performing products and reduce the orders
of products that are not performing well.
Examples of management information systems include;
 Sales management systems – they get input from the point of sale system
 Budgeting systems – gives an overview of how much money is spent within the
organization for the short and long terms.
 Human resource management system – overall welfare of the employees, staff
turnover, etc.
Tactical managers are responsible for the semi-structured decision. MIS systems provide the
information needed to make the structured decision and based on the experience of the tactical
managers, they make judgement calls i.e. predict how much of goods or inventory should be
ordered for the second quarter based on the sales of the first quarter.

Decision Support System (DSS)


Decision support systems are used by senior management to make non-routine decisions.
Decision support systems use input from internal systems (transaction processing systems and
management information systems) and external systems.
The main objective of decision support systems is to provide solutions to problems that are
unique and change frequently. Decision support systems answer questions such as;
 What would be the impact of employees' performance if we double the production lot at
the factory?
 What would happen to our sales if a new competitor entered the market?
Decision support systems use sophisticated mathematical models, and statistical techniques
(probability, predictive modeling, etc.) to provide solutions, and they are very interactive.
Examples of decision support systems include;
 Financial planning systems – it enables managers to evaluate alternative ways of
achieving goals. The objective is to find the optimal way of achieving the goal. For
example, the net profit for a business is calculated using the formula Total Sales less
(Cost of Goods + Expenses). A financial planning system will enable senior executives to
ask what if questions and adjust the values for total sales, the cost of goods, etc. to see the
effect of the decision and on the net profit and find the most optimal way.
 Bank loan management systems – it is used to verify the credit of the loan applicant and
predict the likelihood of the loan being recovered.
Executive Information Systems
Executive Information Systems are strategic-level information systems that are found at the top
of the Pyramid. They help executives and senior managers analyze the environment in which the
organization operates, to identify long-term trends, and to plan appropriate courses of action. The
information in such systems is often weakly structured and comes from both internal and
external sources. Executive Information System are designed to be operated directly by
executives without the need for intermediaries and easily tailored to the preferences of the
individual using them.
Functions of EIS
EIS organizes and presents data and information from both external data sources and internal
MIS or TPS in order to support and extend the inherent capabilities of senior executives.
Functions of a EIS in terms of data processing requirements
Inputs Processing Outputs

External Data Summarizing Summary reports


Internal Files Simulation Forecasts
Pre-defined models "Drilling Down" Graphs / Plots
Some examples of EIS
Executive Information Systems tend to be highly individualized and are often custom made for a
particular client group; however, a number of off-the-shelf EIS packages do exist and many
enterprise level systems offer a customizable EIS module.
The role of EIS
o Are concerned with ease of use
o Are concerned with predicting the future
o Are effectiveness oriented
o Are highly flexible
o Support unstructured decisions
o Use internal and external data sources
o Used only at the most senior management levels

Functional Information Systems


Supporting Business Functions in an Enterprise with Information
The principal business functions in a business firm are:
1. Marketing and sales 3. Accounting and finance
2. Production 4. Human resources
Marketing Information Systems
Marketing activities are directed toward planning, promoting, and selling goods and services to
satisfy the needs of customers and the objectives of the organization.
Marketing information systems support decision making regarding the marketing mix. These
include:
1. Product 3. Place
2. Price 4. Promotion
Sources of Data and Information for Marketing: Boundary-Spanning and Transaction
Processing Subsystems
A marketing information system relies on external information to a far greater degree than other
organizational information systems. It includes two subsystems designed for boundary spanning
- bringing into the firm data and information about the marketplace.
The objective of marketing research is to collect data on the actual customers and the potential
customers, known as prospects. The identification of the needs of the customer is a fundamental
starting point for total quality management (TQM). Electronic commerce on the WEB makes it
easy to compile statistics on actual buyer behaviour.
Marketing research software supports statistical analysis of data. It enables the firm to correlate
buyer behaviour with very detailed geographic variables, demographic variables, and
psychographic variables.
Marketing (competitive) intelligence is responsible for the gathering and interpretation of data
regarding the firm's competitors, and for the dissemination of the competitive information to the
appropriate users. Most of the competitor information comes from corporate annual reports,
media-tracking services, and from reports purchased from external providers, including on-line
database services. The Internet has become a major source of competitive intelligence.
Marketing Mix Subsystems
The marketing mix subsystems support decision making regarding product introduction, pricing,
promotion (advertising and personal selling), and distribution. These decisions are integrated into
the sales forecast and marketing plans against which the ongoing sales results are compared.
Marketing mix subsystems include:
1. Product subsystem 4. Price subsystem
2. Place subsystem 5. Sales forecasting
3. Promotion subsystem
Product Subsystem
The product subsystem helps to plan the introduction of new products. Continually bringing new
products to market is vital in today's competitive environment of rapid change. The product
subsystem should support balancing the degree of risk in the overall new-product portfolio, with
more aggressive competitors assuming higher degrees of risk for a potentially higher payoff.
Although decisions regarding the introduction of new products are unstructured, information
systems support this process in several ways:
1. Professional support systems assist designers in their knowledge work
2. DSSs are used to evaluate proposed new products
3. With a DSS, a marketing manager can score the desirability of a new product.
4. Electronic meeting systems help bring the expertise of people dispersed in space and time to
bear on the problem
5. Information derived from marketing intelligence and research is vital in evaluating new
product ideas.
Place Subsystem
The place subsystem assists the decision makers in making the product available to the customer
at the right place at the right time. The place subsystem helps plan the distribution channels for
the product and track their performance.
The use of information technology has dramatically increased the availability of information on
product movement in the distribution channel. Examples include:
1. Bar-coded Universal Product Code (UPC)
2. Point-of-sale (POS) scanning
3. Electronic data interchange (EDI)
4. Supports just-in-time product delivery and customized delivery
Promotion Subsystem
The promotion subsystem is often the most elaborate in the marketing information system, since
it supports both personal selling and advertising. Media selection packages assist in selecting a
mix of avenues to persuade the potential purchaser, including direct mail, television, print media,
and the electronic media such as the Internet and the WEB in particular. The effectiveness of the
selected media mix is monitored and its composition is continually adjusted.
Database marketing relies on the accumulation and use of extensive databases to segment
potential customers and reach tem with personalized promotional information.
The role of telemarketing, marketing over the telephone, has increased. Telemarketing calls are
well supported by information technology.
Sales management is thoroughly supported with information technology. Customer profitability
analysis help identify high-profit and high-growth customers and target marketing efforts in
order to retain and develop these accounts.
Sales force automation, involves equipping salespeople with portable computers tied into the
corporate information systems. This gives the salespeople instantaneous access to information
and frees them from the reporting paperwork. This increases selling time and the level of
performance. Access to corporate databases is sometimes accompanied by access to corporate
expertise, either by being able to contact the experts or by using expert systems that help specify
the product meeting customer requirements.
Price Subsystem
Pricing decisions find a degree of support from DSSs and access to databases that contain
industry prices. These highly unstructured decisions are made in pursuit of the companys pricing
objectives. General strategies range from profit maximization to forgoing a part of the profit in
order to increase a market share.
Information systems provide an opportunity to finely segment customer groups, and charge
different prices depending on the combination of products and services provided, as well as the
circumstances of the sale transaction.
Sales Forecasting
Based on the planned marketing mix and outstanding orders, sales are forecast and a full
marketing plan is developed. Sale forecasting is an area where any quantitative methods
employed must be tempered with human insight and experience. The actual sales will depend to
a large degree on the dynamics of the environment.
Qualitative techniques are generally used for environmental forecasting - an attempt to predict
the social, economic, legal, and technological environment in which the company will try to
realize its plans. Sales forecasting uses numerous techniques, which include:
1. Group decision making techniques are used to elicit broad expert opinion
2. Scenario analysis in which each scenario in this process is a plausible future environment
3. Extrapolation of trends and cycles through a time-series analysis.
Manufacturing or Production Information Systems
Global competitive pressures of the information society have been highly pronounced in
manufacturing and have radically changed it. The new marketplace calls for manufacturing that
are:
1. Lean - highly efficient, using fewer input resources in production through better engineering
and through production processes that rely on low inventories and result in less waste.
2. Agile - fit for time-based competition. Both the new product design and order fulfilment are
drastically shortened.
3. Flexible - able to adjust the product to a customer's preferences rapidly and cost effectively.
4. Managed for quality - by measuring quality throughout the production process and following
world standards, manufacturers treat quality as a necessity and not a high-price option.
Structure of Manufacturing Information Systems
Information technology must play a vital role in the design and manufacturing processes.
Manufacturing information systems are among the most difficult both to develop and to
implement.
TPSs are embedded in the production process or in other company processes. The data provided
by the transaction processing systems are used by management support subsystems, which are
tightly integrated and interdependent.
Manufacturing information subsystems include:
1. Product design and engineering
2. Product scheduling
3. Quality control
4. Facilities planning, production costing, logistics and inventory subsystems
Product Design and Engineering
Product design and engineering are widely supported today by computer-aided design (CAD)
and computer-aided engineering (CAE) systems. CAD systems assist the designer with
automatic calculations and display of surfaces while storing the design information in databases.
The produced designs are subject to processing with CAE systems to ensure their quality, safety,
manufacturability, and cost-effectiveness. CAD/CAE systems increasingly eliminate paperwork
from the design process, while speeding up the process itself. As well, the combined techniques
of CAD/CAE and rapid prototyping cut time to market.
Product Scheduling
Production scheduling is the heart of the manufacturing information system. This complex
subsystem has to ensure that an appropriate combination of human, machinery, and material
resources will be provided at an appropriate time in order to manufacture the goods.
Production scheduling and the ancillary processes are today frequently controlled with
a manufacturing resource planning system as the main informational tool. This elaborate
software converts the sales forecast for the plants products into a detailed production plan and
further into a master schedule of production.
Computer integrated manufacturing (CIM) is a strategy through which a manufacturer takes
control of the entire manufacturing process. The process starts with CAD and CAE and
continues on the factory floor where robots and numerically controlled machinery are installed -
and thus computer-aided manufacturing (CAM) is implemented. A manufacturing system based
on this concept can turn out very small batches of a particular product as cost-effectively as a
traditional production line can turn out millions of identical products. A full-fledged CIM is
extremely difficult to implement; indeed, many firms have failed in their attempts to do so.
Quality Control
The quality control subsystem of a manufacturing information system relies on the data collected
on the shop floor by the sensors embedded in the process control systems.
Total quality management (TQM) is a management technique for continuously improving the
performance of all members and units of a firm to ensure customer satisfaction. In particular, the
principles of TQM state that quality comes from improving the design and manufacturing
process, rather than Ainspecting out@ defective products. The foundation of quality is also
understanding and reducing variation in the overall manufacturing process.
Facilities Planning, Production Costing, Logistics and Inventory Subsystems
Among the higher-level decision making supported by manufacturing information systems are
facilities planning - locating the sites for manufacturing plants, deciding on their production
capacities, and laying out the plant floors.
Manufacturing management requires a cost control program, relying on the information systems.
Among the informational outputs of the production costing subsystem are labor and equipment
productivity reports, performance of plants as cost centers, and schedules for equipment
maintenance and replacement.
Managing the raw-materials, packaging, and the work in progress inventory is a responsibility of
the manufacturing function. In some cases, inventory management is combined with the general
logistics systems, which plan and control the arrival of purchased goods into the firm as well as
shipments to the customers.
Accounting and Financial Information Systems
The financial function of the enterprise consists in taking stock of the flows of money and other
assets into and out of an organization, ensuring that its available resources are properly used and
that the organization is financially fit. The components of the accounting system include:
1. Accounts receivable records 4. Inventory control records
2. Accounts payable records 5. General ledgers
3. Payroll records
Financial information systems rely on external sources, such as on-line databases and custom
produced reports, particularly in the areas of financial forecasting and funds management. The
essential functions that financial information systems perform include:
1. Financial forecasting and planning 3. Funds management
2. Financial control 4. Internal auditing
Financial Forecasting
Financial forecasting is the process of predicting the inflows of funds into the company and the
outflows of funds from it for a long term into the future. Outflows of funds must be balanced
over the long term with the inflows. With the globalization of business, the function of financial
forecasting has become more complex, since the activities in multiple national markets have to
be consolidated, taking into consideration the vagaries of multiple national currencies. Scenario
analysis is frequently employed in order to prepare the firm for various contingencies.
Financial forecasts are based on computerized models known as cash-flow models. They range
from rather simple spreadsheet templates to sophisticated models developed for the given
industry and customized for the firm or, in the case of large corporations to specify modeling of
their financial operations. Financial forecasting serves to identify the need for funds and their
sources.
Financial Control
The primary tools of financial control are budgets. A budget specifies the resources committed to
a plan for a given project or time period. Fixed budgets are independent of the level of activity of
the unit for which the budget is drawn up. Flexible budgets commit resources depending on the
level of activity.
Spreadsheet programs are the main budgeting tools. Spreadsheets are the personal productivity
tools in use today in budget preparation.
In the systems-theoretic view, budgets serve as the standard against which managers can
compare the actual results by using information systems. Performance reports are used to
monitor budgets of various managerial levels. A performance report states the actual financial
results achieved by the unit and compares them with the planned results.
Along with budgets and performance reports, financial control employs a number of financial
ratios indicating the performance of the business unit. A widely employed financial ratio
is return on investment (ROI). ROS shows how well a business unit uses its resources. Its value
is obtained by dividing the earnings of the business unit by its total assets.
Funds Management
Financial information systems help to manage the organization's liquid assets, such as cash or
securities, for high yields with the lowest degree of loss risk. Some firms deploy computerized
systems to manage their securities portfolios and automatically generate buy or sell orders.
Internal Auditing
The audit function provides an independent appraisal of an organization's accounting, financial,
and operational procedures and information. All large firms have internal auditors, answerable
only to the audit committee of the board of directors. The staff of the chief financial officer of
the company performs financial and operational audits. During a financial audit, an appraisal is
made of the reliability and integrity of the company's financial information and of the means
used to process it. An operational audit is an appraisal of how well management utilizes
company resources and how well corporate plans are being carried out.
Human Resource Information Systems
A human resource information system (HRIS) supports the human resources function of an
organization with information. The name of this function reflects the recognition that people who
work in a firm are frequently its most valuable resources. The complexity of human resource
management has grown immensely over recent years, primary due to the need to conform with
new laws and regulations.
A HRIS has to ensure the appropriate degree of access to a great variety of internal stakeholders,
including:
1. The employees of the Human Resources department in performance of their duties
2. All the employees of the firm wishing ti inspect their own records
3. All the employees of the firm seeking information regarding open positions or available
benefit plans
4. Employees availing themselves of the computer-assisted training and evaluation opportunities
5. Managers throughout the firm in the process of evaluating their subordinates and making
personnel decisions
6. Corporate executives involved in tactical and strategic planning and control
Transaction Processing Subsystems and Databases of Human Resource Information
Systems
At the heart of HRIS are its databases, which are in some cases integrated into a single human
resource database. The record of each employee in a sophisticated employee database may
contain 150 to 200 data items, including the personal data, educational history and skills,
occupational background, and the history of occupied positions, salary, and performance in the
firm. Richer multimedia databases are not assembled by some firms in order to facilitate fast
formation of compatible teams of people with complementary skills.
Other HRIS databases include:
1. Applicant databases 4. Benefit databases
2. Position inventory 5. External databases
3. Skills inventory
Information Subsystems for Human Resource Management
The information subsystems of HRIS reflect the flow of human resources through the firm, from
planning and recruitment to termination. A sophisticated HRIS includes the following
subsystems:
1. Human resource planning
2. Recruiting and workforce management
3. Compensation and benefits
4. Government reporting and labour relations support

Human Resource Planning


To identify the human resources necessary to accomplish the long-term objectives of a firm, we
need to project the skills, knowledge, and experience of the future employees.
Recruiting and Workforce Management
Based on the long-term resource plan, a recruitment plan is developed. The plan lists the
currently unfilled positions and those expected to become vacant due to turnover.
The life-cycle transitions of the firm's workforce - hiring, promotion and transfer, and
termination - have to be supported with the appropriate information system components.
Compensation and Benefits
Two principal external stakeholders have an abiding interest in the human resource policies of
organizations. These are:
1. Various levels of government
2. Labor unions
Integrating Functional Systems for Superior Organizational Performance
Functional information systems rarely stand alone. This reflects the fact that the functions they
support should, as much as possible, connect with each other seamlessly in order to serve the
firms customers. Customers expect timely order delivery, often on a just-in-time schedule;
quality inspection to their own standards; flexible credit terms; post-delivery service; and often,
participation in the product design process.
Information technology provides vital support for integrating internal business processes, cutting
across functional lines, and for integrating operations with the firm's business partners, its
customers and suppliers.
Decision Support Systems (DSS)
Decision support systems (DSS) are interactive software-based systems intended to help
managers in decision-making by accessing large volumes of information generated from various
related information systems involved in organizational business processes, such as office
automation system, transaction processing system, etc.
DSS uses the summary information, exceptions, patterns, and trends using the analytical
models. A decision support system helps in decision-making but does not necessarily give a
decision itself. The decision makers compile useful information from raw data, documents,
personal knowledge, and/or business models to identify and solve problems and make
decisions.
Programmed and Non-programmed Decisions
There are two types of decisions –
Programmed and Non-programmed Decisions.
Programmed decisions are basically automated processes, general routine work, where −
 These decisions have been taken several times.
 These decisions follow some guidelines or rules.
For example, selecting a reorder level for inventories, is a programmed decision.
Non-programmed decisions occur in unusual and non-addressed situations, so −
 It would be a new decision.
 There will not be any rules to follow.
 These decisions are made based on the available information.
 These decisions are based on the manger's discretion, instinct, perception and judgment.
For example, investing in a new technology is a non-programmed decision.
Decision support systems generally involve non-programmed decisions. Therefore, there will be
no exact report, content, or format for these systems. Reports are generated on the fly.
Attributes of a DSS
 Adaptability and flexibility  Ease of development
 High level of Interactivity  Extendibility
 Ease of use  Support for modeling and analysis
 Efficiency and effectiveness  Support for data access
 Complete control by decision-  Standalone, integrated, and Web-
makers based
Characteristics of a DSS
 Support for decision-makers in semi-structured and unstructured problems.
 Support for managers at various managerial levels, ranging from top executive to line
managers.
 Support for individuals and groups. Less structured problems often requires the
involvement of several individuals from different departments and organization level.
 Support for interdependent or sequential decisions.
 Support for intelligence, design, choice, and implementation.
 Support for variety of decision processes and styles.
 DSSs are adaptive over time.
Benefits of DSS
 Improves efficiency and speed of decision-making activities.
 Increases the control, competitiveness and capability of futuristic decision-making of the
organization.
 Facilitates interpersonal communication.
 Encourages learning or training.
 Since it is mostly used in non-programmed decisions, it reveals new approaches and sets
up new evidences for an unusual decision.
 Helps automate managerial processes.
Components of a DSS
Following are the components of the Decision Support System −
 Database Management System (DBMS) − To solve a problem the necessary data may
come from internal or external database. In an organization, internal data are generated
by a system such as TPS and MIS. External data come from a variety of sources such as
newspapers, online data services, databases (financial, marketing, human resources).
 Model Management System − It stores and accesses models that managers use to make
decisions. Such models are used for designing manufacturing facility, analyzing the
financial health of an organization, forecasting demand of a product or service, etc.
 Support Tools − Support tools like online help; pulls down menus, user interfaces,
graphical analysis, error correction mechanism, facilitates the user interactions with the
system.
Classification of DSS
There are several ways to classify DSS. Hoi Apple and Whinstone classifies DSS as follows −
 Text Oriented DSS − It contains textually represented information that could have a
bearing on decision. It allows documents to be electronically created, revised and
viewed as needed.
 Database Oriented DSS − Database plays a major role here; it contains organized and
highly structured data.
 Spreadsheet Oriented DSS − It contains information in spread sheets that allows create,
view, modify procedural knowledge and also instructs the system to execute self-
contained instructions. The most popular tool is Excel and Lotus 1-2-3.
 Solver Oriented DSS − It is based on a solver, which is an algorithm or procedure
written for performing certain calculations and particular program type.
 Rules Oriented DSS − It follows certain procedures adopted as rules.
 Rules Oriented DSS − Procedures are adopted in rules oriented DSS. Export system is
the example.
 Compound DSS − It is built by using two or more of the five structures explained
above.
Types of DSS
Following are some typical DSSs −
 Status Inquiry System − It helps in taking operational, management level, or middle
level management decisions, for example daily schedules of jobs to machines or
machines to operators.
 Data Analysis System − It needs comparative analysis and makes use of formula or an
algorithm, for example cash flow analysis, inventory analysis etc.
 Information Analysis System − In this system data is analyzed and the information
report is generated. For example, sales analysis, accounts receivable systems, market
analysis etc.
 Accounting System − It keeps track of accounting and finance related information, for
example, final account, accounts receivables, accounts payables, etc. that keep track of
the major aspects of the business.
 Model Based System − Simulation models or optimization models used for decision-
making are used infrequently and creates general guidelines for operation or
management.
Executive Information System (EIS)
Executive Information Systems are intended to be used by the senior managers directly to
provide support to non-programmed decisions in strategic management.
These information are often external, unstructured and even uncertain. Exact scope and context
of such information is often not known beforehand.
This information is intelligence based −
 Market intelligence
 Investment intelligence
 Technology intelligence
Features of Executive Information System

Examples of Intelligent Information


Following are some examples of intelligent information, which is often the source of an EIS −
 External databases
 Technology reports like patent records etc.
 Technical reports from consultants
 Market reports
 Confidential information about competitors
 Speculative information like market conditions
 Government policies
 Financial reports and information
Advantages of EIS
 Easy for upper level executive to use
 Ability to analyze trends
 Augmentation of managers' leadership capabilities
 Enhance personal thinking and decision-making
 Contribution to strategic control flexibility
 Enhance organizational competitiveness in the market place
 Instruments of change
 Increased executive time horizons.
 Better reporting system
 Improved mental model of business executive
 Help improve consensus building and communication
 Improve office automation
 Reduce time for finding information
 Early identification of company performance
 Detail examination of critical success factor
 Better understanding
 Time management
 Increased communication capacity and quality
Disadvantage of EIS
 Functions are limited
 Hard to quantify benefits
 Executive may encounter information overload
 System may become slow
 Difficult to keep current data
 May lead to less reliable and insecure data
 Excessive cost for small company

Knowledge Management System (KMS)


All the systems we are discussing here come under knowledge management category. A
knowledge management system is not radically different from all these information systems, but
it just extends the already existing systems by assimilating more information.
As we have seen, data is raw facts, information is processed and/or interpreted data, and
knowledge is personalized information.
What is Knowledge?
 Personalized information
 State of knowing and understanding
 An object to be stored and manipulated
 A process of applying expertise
 A condition of access to information
 Potential to influence action
Sources of Knowledge of an Organization
 Intranet
 Data warehouses and knowledge repositories
 Decision support tools
 Groupware for supporting collaboration
 Networks of knowledge workers
 Internal expertise
Definition of KMS
A knowledge management system comprises a range of practices used in an organization to
identify, create, represent, distribute, and enable adoption to insight and experience. Such
insights and experience comprise knowledge, either embodied in individual or embedded in
organizational processes and practices.
Purpose of KMS
 Improved performance
 Competitive advantage
 Innovation
 Sharing of knowledge
 Integration
 Continuous improvement by −
o Driving strategy
o Starting new lines of business
o Solving problems faster
o Developing professional skills
o Recruit and retain talent
Activities in Knowledge Management
 Start with the business problem and the business value to be delivered first.
 Identify what kind of strategy to pursue to deliver this value and address the KM
problem.
 Think about the system required from a people and process point of view.
 Finally, think about what kind of technical infrastructure are required to support the
people and processes.
 Implement system and processes with appropriate change management and iterative
staged release.
Level of Knowledge Management
Business Intelligence System (BIS)
The term 'Business Intelligence' has evolved from the decision support systems and gained
strength with the technology and applications like data warehouses, Executive Information
Systems and Online Analytical Processing (OLAP).
Business Intelligence System is basically a system used for finding patterns from existing data
from operations.
Characteristics of BIS
 It is created by procuring data and information for use in decision-making.
 It is a combination of skills, processes, technologies, applications and practices.
 It contains background data along with the reporting tools.
 It is a combination of a set of concepts and methods strengthened by fact-based support
systems.
 It is an extension of Executive Support System or Executive Information System.
 It collects, integrates, stores, analyzes, and provides access to business information
 It is an environment in which business users get reliable, secure, consistent,
comprehensible, easily manipulated and timely information.
 It provides business insights that lead to better, faster, more relevant decisions.
Benefits of BIS
 Improved Management Processes.
 Planning, controlling, measuring and/or applying changes that results in increased
revenues and reduced costs.
 Improved business operations.
 Fraud detection, order processing, purchasing that results in increased revenues and
reduced costs.
 Intelligent prediction of future.
Approaches of BIS
For most companies, it is not possible to implement a proactive business intelligence system at
one go. The following techniques and methodologies could be taken as approaches to BIS −
 Improving reporting and analytical capabilities
 Using scorecards and dashboards
 Enterprise Reporting
 On-line Analytical Processing (OLAP) Analysis
 Advanced and Predictive Analysis
 Alerts and Proactive Notification
 Automated generation of reports with user subscriptions and "alerts" to problems and/or
opportunities.
Capabilities of BIS
 Data Storage and Management − o Managed reporting
o Visualization
o Data ware house
o Scorecard
o Ad hoc analysis
 Query, Reporting and Analysis
o Data quality
o Data mining o Ad hoc Analysis
 Information Delivery o Production reporting
o OLAP analysis
o Dashboard
o Collaboration /search

Geographic Information System (GIS)


A Geographic Information System (GIS) is a computer system that analyzes and displays
geographically referenced information. It uses data that is attached to a unique location.
Definition
A geographic information system (GIS) is a system designed to capture, store, manipulate,
analyze, manage, and present all types of geographical data. The key word to this technology
is Geography – this means that some portion of the data is spatial. In other words, data that is in
some way referenced to locations on the earth.
Coupled with this data is usually tabular data known as attribute data. Attribute data can be
generally defined as additional information about each of the spatial features. An example of
this would be schools. The actual location of the schools is the spatial data. Additional data such
as the school name, level of education taught, student capacity would make up the attribute data.
It is the partnership of these two data types that enables GIS to be such an effective problem
solving tool through spatial analysis.

GIS is more than just software. People and methods are combined with geospatial software and
tools, to enable spatial analysis, manage large datasets, and display information in a
map/graphical form.
A geographic information system (GIS) is a conceptualized framework that provides the
ability to capture and analyze spatial and geographic data. GIS applications (or GIS apps) are
computer-based tools that allow the user to create interactive queries (user-created searches),
store and edit spatial and non-spatial data, analyze spatial information output, and visually share
the results of these operations by presenting them as maps.
Geographic information systems are utilized in multiple technologies, processes, techniques and
methods. They are attached to various operations and numerous applications, that relate to:
engineering, planning, management, transport/logistics, insurance, telecommunications, and
business. For this reason, GIS and location intelligence applications are at the foundation of
location-enabled services, that rely on geographic analysis and visualization.
GIS provides the capability to relate begun to open new avenues of scientific
previously unrelated information, through inquiry and studies.
the use of location as the "key index
variable". Locations and extents that are
found in the Earth's spacetime, are able to be
recorded through the date and time of
occurrence, along with x, y, and
z coordinates;
representing, longitude (x), latitude (y),
and elevation (z). All Earth-based, spatial–
temporal, location and extent references,
should be relatable to one another, and
ultimately, to a "real" physical location or
extent. This key characteristic of GIS, has

International Information Systems


An international information systems architecture consists of the basic information systems
required by organizations to coordinate worldwide trade and other activities. The basic strategy
to follow when building an international system is to understand the global environment in which
your firm is operating. This means understanding the overall market forces, or business drivers,
that are pushing your industry toward global competition.

A business driver is a force in the


environment to which businesses must
respond and that influences the direction of
the business.

The Global Environment: Business Drivers and Challenges


The global business drivers can be divided into two groups:
General cultural factors and Specific business factors.
Easily recognized general cultural factors have driven internationalization since World War II.
The development of global communications has created a global village in a second sense: A
global culture created by television, the Internet, and other globally shared media such as movies
now permits different cultures and peoples to develop common expectations about right and
wrong, desirable and undesirable, heroic and cowardly.
Responding to demand, global production and operations have emerged with precise online
coordination between far-flung production facilities and central headquarters thousands of miles
away. The new global markets and pressure toward global production and operation have called
forth whole new capabilities for global coordination.
Finally, global markets, production, and administration create the conditions for powerful,
sustained global economies for scale. Not all industries are similarly affected by these trends.
Clearly, manufacturing has been much more affected than services that still tend to be domestic
and highly inefficient. However, the localism of services is breaking down in
telecommunications, entertainment, transportation, finance law, and general business.
Business Challenges
As a cultural level, particularism, making judgments and taking action on the basis of narrow or
personal characteristics, in all its forms (religious, nationalistics, ethnic, regionalism, geopolitical
position) rejects the very concept of a shared global culture and rejects the penetration of
domestic markets by foreign goods and services.
Transborder data flow is defined as the movement of information across international
boundaries in any form.
State of The Art
There are significant difficulties in building appropriate international architectures. The
difficulties involve planning a system appropriate to the firm's global strategy, structuring the
organization of systems and business units, solving implementation issues, and choosing the
right technical platform.
Organizing International Information Systems
Three organizational issues face corporations seeking a global position: choosing a strategy,
organizing the business, and organizing the systems management area.
Unit II System Analysis and Design
Syllabus: System development methodologies, Systems Analysis and Design, Data flow
Diagram (DFD), Decision table, Entity Relationship (ER), Object Oriented Analysis and Design
(OOAD), UML diagram.

System Development Methodologies


A system development methodology refers to the framework that is used to structure, plan, and
control the process of developing an information system. A wide variety of such frameworks
have evolved over the years, each with its own recognized strengths and weaknesses. One system
development methodology is not necessarily suitable for use by all projects. Each of the
available methodologies is best suited to specific kinds of projects, based on various technical,
organizational, project and team considerations.
Basic Principles:
Project is divided into sequential phases, with some overlap and splash back acceptable
between phases.
Emphasis is on planning, time schedules, target dates, budgets and implementation of an
entire system at one time.
Tight control is maintained over the life of the project through the use of extensive written
documentation, as well as through formal reviews and approval/signoff by the user and
information technology management occurring at the end of most phases before beginning
the next phase.
Systems Development Life Cycle (SDLC)
A structured step-by-step approach for developing an information system. A sequence of events
carried out by analysts, designers and users in the development and implementation of
Information system. One of the basic notions of the system development process is SDLC
models which stands for System Development Life Cycle models. SDLC – is a continuous
process, which starts from the moment, when it’s made a decision to launch the project, and it
ends at the moment of its full remove from the exploitation.
Typical activities in SDLC include:
The activities involved in each SDLC phase are:

SDLC PHASE ACTIVITIES

1. Preliminary Investigation / Feasibility • Feasibility Study


Study
• Define the system to be developed
(Planning)
• Set the project scope
• Develop the project plan

2. Determination of System Requirement • Gather business requirements


(Analysis)

3. Designing the System • Design the technical architecture


• Design system models

4. Development of System • Build technical architecture


• Build databases and programs

5. System Testing • Write test conditions


• Perform testing

6. System Implementation • Write user documentation


• Provide training

7. System Maintenance • Build a help desk


• Support system changes

SDLC MODELS / Methodologies


There is no one single SDLC model. They are divided into main groups, each with its features
and weaknesses.
Evolving from the first and oldest “waterfall” SDLC model, their variety significantly expanded.
The SDLC models diversity is predetermined by the wide number of product types – starting
with a web application development to a complex medical software. And if you take one of the
SDLC models mentioned below as the basis – in any case, it should be adjusted to the features of
the product, project, and company. The most used, popular and important SDLC models are
given below:
Types of System development life cycles (SDLC)
1. Waterfall Model
2. Iterative and Incremental Method
3. Spiral Method (SDM)
4. V-Shaped Model
5. Agile Model
6. Prototyping Model
1. Waterfall SDLC Model
Waterfall – is a cascade SDLC model, in which development process looks like the flow, moving
step by step through the phases of analysis, projecting, realization, testing, implementation, and
support. This SDLC model includes gradual execution of every stage completely. This process is
strictly documented and predefined with features expected to every phase of this system
development life cycle model.

ADVANTAGES DISADVANTAGES

Simple to use and understand The software is ready only after the last stage is over

Management simplicity thanks to its rigidity: every High risks and uncertainty
phase has a defined result and process review

Development stages go one by one Not the best choice for complex and object-oriented projects
ADVANTAGES DISADVANTAGES

Perfect for the small or mid-sized projects where Inappropriate for the long-term projects
requirements are clear and not equivocal

Easy to determine the key points in the development The progress of the stage is hard to measure while it is still in
cycle the development

Easy to classify and prioritize tasks Integration is done at the very end, which does not give the
option of identifying the problem in advance

Use cases for the Waterfall SDLC model:


 The requirements are precisely documented
 Product definition is stable
 The technologies stack is predefined which makes it not dynamic
 No ambiguous requirements
 The project is short
2. Iterative SDLC Model
The Iterative SDLC model does not need the full list of requirements before the project starts.
The development process may start with the requirements to the functional part, which can be
expanded later. The process is repetitive, allowing to make new versions of the product for every
cycle. Every iteration (which last from two to six weeks) includes the development of a separate
component of the system, and after that, this component is added to the functional developed
earlier. Speaking with math terminology, the iterative model is a realization of the sequential
approximation method; that means a gradual closeness to the planned final product shape.
ADVANTAGES DISADVANTAGES

Some functions can be quickly developed at the Iterative model requires more resources than the waterfall
beginning of the development lifecycle model

The paralleled development can be applied Constant management is required

The progress is easy measurable Issues with architecture or design may occur because not all
the requirements are foreseen during the short planning stage

The shorter iteration is - the easier testing and Bad choice for the small projects
debugging stages are

It is easier to control the risks as high-risk tasks are The process is difficult to manage
completed first

Problems and risks defined within one iteration can be The risks may not be completely determined even at the final
prevented in the next sprints stage of the project

Flexibility and readiness to the changes in the Risks analysis requires involvement of the highly-qualified
requirements specialists

Use cases for the Iteration model:


 The requirements to the final product are strictly predefined
 Applied to the large-scale projects
 The main task is predefined, but the details may advance with the time

3. Spiral SDLC Model


Spiral model – is SDLC model, which
combines architecture and prototyping by
stages. It is a combination of the Iterative
and Waterfall SDLC models with the
significant accent on the risk analysis. The
main issue of the spiral model – is defining
the right moment to make a step into the
next stage. The preliminary set time frames
are recommended as the solution to this
issue. The shift to the next stage is done
according to the plan, even if the work on
the previous stage isn’t done yet. The plan is
introduced basing on the statistic data,
received during the previous projects even
from the personal developer’s experience.
ADVANTAGES DISADVANTAGES

Lifecycle is divided into small parts, and if the risk Can be quite expensive
concentration is higher, the phase can be finished
earlier to address the treats

The development process is precisely documented yet The risk control demands involvement of the highly-
scalable to the changes skilled professionals

The scalability allows to make changes and add new Can be ineffective for the small projects
functionality even at the relatively late stages

The earlier working prototype is done - sooner users Big number of the intermediate stages requires
can point out the flaws excessive documentation

Use cases for the Spiral model


 Customer isn’t sure about the requirements
 Major edits are expected during the development cycle
 The projects with mid or high-level risk, where it is important to prevent these risks
 The new product that should be released in a few stages to have enough of clients
feedback

4. V-shaped SDLC Model

V-shaped SDLC model is an expansion of


classic waterfall model and it’s based on
associated test stage for the every
development stage. This is a very strict
model and the next stage is started only after
the previous phase. This is also called
“Validation and verification” model. Every
stage has the current process control, to
make sure that the conversion to the next
stage is possible.
ADVANTAGES DISADVANTAGES

Every stage of V-shaped model has strict results so it’s easy to Lack of the flexibility
control

Testing and verification take place in the early stages Bad choice for the small projects

Good for the small projects, where requirements are static and clear Relatively big risks

Use cases for the V-shaped model:


 For the projects where an accurate product testing is required
 For the small and mid-sized projects, where requirements are strictly predefined
 The engineers of the required qualification, especially testers, are within easy reach.

5. Agile SDLC Model


In the agile methodology after every development iteration, the customer is able to see the result
and understand if he is satisfied with it or he is not. This is one of the advantages of the agile
software development life cycle model. One of its disadvantages is that with the absence of
defined requirements it is difficult to estimate the resources and development cost. Extreme
programming is one of the practical use of the agile model. The basis of such model consists of
short weekly meetings – Sprints which are the part of the Scrum approach.
ADVANTAGES DISADVANTAGES

Corrections of functional requirements are implemented into Difficulties with measuring the final cost because of
the development process to provide the competitiveness permanent changes

Project is divided by short and transparent iterations The team should be highly professional and client-
oriented

Risks are minimized thanks to the flexible change process New requirements may conflict with the existing
architecture

Fast release of the first product version With all the corrections and changes there is possibility
that the project will exceed expected time

Use cases for the Agile model:


 The users’ needs change dynamically
 Less price for the changes implemented because of the many iterations
 Unlike the Waterfall model, it requires only initial planning to start the project

6. Prototyping Model
It refers to the activity of creating prototypes of software applications, for example, incomplete
versions of the software program being developed. It is an activity that can occur in software
development and It used to visualize some component of the software to limit the gap of
misunderstanding the customer requirements by the development team. This also will reduce the
iterations may occur in the waterfall approach and hard to be implemented due to the inflexibility
of the waterfall approach. So, when the final prototype is developed, the requirement is
considered to be frozen.

It has some types, such as:


 Throwaway prototyping: Prototypes that are eventually discarded rather than becoming
a part of the finally delivered software
 Evolutionary prototyping: prototypes that evolve into the final system through an
iterative incorporation of user feedback.
 Incremental prototyping: The final product is built as separate prototypes. In the end,
the separate prototypes are merged in an overall design.
 Extreme prototyping: used in web applications mainly. Basically, it breaks down web
development into three phases, each one based on the preceding one. The first phase is a
static prototype that consists mainly of HTML pages. In the second phase, the screens are
programmed and fully functional using a simulated services layer. In the third phase, the
services are implemented

The usage

This process can be used with any software developing life cycle model. While this shall be
chosen when you are developing a system has user interactions. So, if the system does not
have user interactions, such as a system does some calculations shall not have prototypes.

Advantages Disadvantages
 Insufficient analysis. User confusion of prototype
 Reduced time and costs, but this can be a and finished system.
disadvantage if the developer loses time in  Developer misunderstanding of user objectives.
developing the prototypes.  Excessive development time of the prototype.
 Improved and increased user involvement.  It is costly to implement the prototypes

Systems Analysis and Design


Systems development is systematic process which includes phases such as planning, analysis,
design, deployment, and maintenance.
Here, we will primarily focus on −
 Systems analysis  Systems design
Systems Analysis
It is a process of collecting and interpreting facts, identifying the problems, and decomposition
of a system into its components.
System analysis is conducted for the purpose of studying a system or its parts in order to
identify its objectives. It is a problem solving technique that improves the system and ensures
that all the components of the system work efficiently to accomplish their purpose.
Analysis specifies what the system should do.
Systems Design
It is a process of planning a new business system or replacing an existing system by defining its
components or modules to satisfy the specific requirements. Before planning, you need to
understand the old system thoroughly and determine how computers can best be used in order to
operate efficiently.
System Design focuses on how to accomplish the objective of the system.
System Analysis and Design (SAD) mainly focuses on −
 Systems
 Processes
 Technology
An effective System Development Life Cycle (SDLC) should result in a high quality system
that meets customer expectations, reaches completion within time and cost evaluations, and
works effectively and efficiently in the current and planned Information Technology
infrastructure.
System Development Life Cycle (SDLC) is a conceptual model which includes policies and
procedures for developing or altering systems throughout their life cycles.
SDLC is used by analysts to develop an information system. SDLC includes the following
activities −
 requirements  deployment
 design  operations
 implementation  maintenance
 testing
Phases of SDLC
Systems Development Life Cycle is a systematic approach which explicitly breaks down the
work into phases that are required to implement either new or modified Information System.

Feasibility Study or Planning


 Define the problem and scope of existing system.
 Overview the new system and determine its objectives.
 Confirm project feasibility and produce the project Schedule.
 During this phase, threats, constraints, integration and security of system are also
considered.
 A feasibility report for the entire project is created at the end of this phase.
Analysis and Specification
 Gather, analyze, and validate the information.
 Define the requirements and prototypes for new system.
 Evaluate the alternatives and prioritize the requirements.
 Examine the information needs of end-user and enhances the system goal.
 A Software Requirement Specification (SRS) document, which specifies the software,
hardware, functional, and network requirements of the system is prepared at the end of this
phase.
System Design
 Includes the design of application, network, databases, user interfaces, and system interfaces.
 Transform the SRS document into logical structure, which contains detailed and complete set of
specifications that can be implemented in a programming language.
 Create a contingency, training, maintenance, and operation plan.
 Review the proposed design. Ensure that the final design must meet the requirements stated in SRS
document.
 Finally, prepare a design document which will be used during next phases.
Implementation
 Implement the design into source code through coding.
 Combine all the modules together into training environment that detects errors and defects.
 A test report which contains errors is prepared through test plan that includes test related tasks such as
test case generation, testing criteria, and resource allocation for testing.
 Integrate the information system into its environment and install the new system.
Maintenance/Support
 Include all the activities such as phone support or physical on-site support for users that is required
once the system is installing.
 Implement the changes that software might undergo over a period of time, or implement any new
requirements after the software is deployed at the customer location.
 It also includes handling the residual errors and resolve any issues that may exist in the system even
after the testing phase.
 Maintenance and support may be needed for a longer time for large systems and for a short time for
smaller systems.
Life Cycle of System Analysis and Design
The following diagram shows the complete
life cycle of the system during analysis and
design phase.

Role of System Analyst


The system analyst is a person who is thoroughly aware of the system and guides the system
development project by giving proper directions. He is an expert having technical and
interpersonal skills to carry out development tasks required at each phase.
He pursues to match the objectives of information system with the organization goal.
Main Roles
 Defining and understanding the requirement of user through various Fact finding techniques.
 Prioritizing the requirements by obtaining user consensus.
 Gathering the facts or information and acquires the opinions of users.
 Maintains analysis and evaluation to arrive at appropriate system which is more user friendly.
 Suggests many flexible alternative solutions, pick the best solution, and quantify cost and benefits.
 Draw certain specifications which are easily understood by users and programmer in precise and
detailed form.
 Implemented the logical design of system which must be modular.
 Plan the periodicity for evaluation after it has been used for some time, and modify the system as
needed.

Attributes of a Systems Analyst

The following figure shows the attributes a


systems analyst should possess −

 Curiosity to learn about new organization


Management Skills
Interpersonal Skills
 Understand users jargon and practices.
 Interface with users and programmer.  Resource & project management.
 Facilitate groups and lead smaller teams.  Change & risk management.
 Managing expectations.  Understand the management functions
 Good understanding, communication, thoroughly.
selling and teaching abilities.
 Motivator having the confidence to solve
queries.
Technical Skills
Analytical Skills  Knowledge of computers and software.
 Keep abreast of modern development.
 System study and organizational  Know of system design tools.
knowledge  Breadth knowledge about new
 Problem identification, problem analysis, technologies.
and problem solving
 Sound commonsense
 Ability to access trade-off
System Flowcharts
A system flowchart is a visual representation of processes, decisions, inputs and outputs that
together form a system.
What is a system flowchart?

System flowcharts are a way of displaying To illustrate this, symbols are used. They are
how data flows in a system and how connected together to show what happens to
decisions are made to control events. data and where it goes.
Note that system flow charts are very similar
to data flow charts. Data flow charts do not
include decisions, they just show the path
that data takes, where it is held, processed,
and then output.

Symbols used in flowchart

Using system flowchart ideas


This system flowchart is a diagram for a
'cruise control' for a car. The cruise control
keeps the car at a steady speed that has been
set by the driver.

The flowchart shows what the outcome is if the car is going too fast or too slow. The system is
designed to add fuel, or take it away and so keep the car's speed constant. The output (the car's
new speed) is then fed back into the system via the speed sensor.

Other examples of uses for system diagrams include:


 aircraft control
 central heating
 automatic washing machines
 booking systems for airlines

Decision Tables
A Decision table represents conditions and the respective actions to be taken to address them, in
a structured tabular format.
It is a powerful tool to debug and prevent errors. It helps group similar information into a single
table and then by combining tables it delivers easy and convenient decision-making.
Decision tables are a method of describing the complex logical relationship in a precise manner
which is easily understandable.
 It is useful in situations where the resulting actions depend on the occurrence of one or
several combinations of independent conditions.
 It is a matrix containing row or columns for defining a problem and the actions.

Creating Decision Table

To create the decision table, the developer must follow basic four steps:

 Identify all possible conditions to be addressed


 Determine actions for all identified conditions
 Create Maximum possible rules
 Define action for each rule
Decision Tables should be verified by end-users and can lately be simplified by eliminating
duplicate rules and actions.

Components of a Decision Table

 Condition Stub − It is in the upper left quadrant which lists all the condition to be
checked.
 Action Stub − It is in the lower left quadrant which outlines all the action to be carried
out to meet such condition.
 Condition Entry − It is in upper right quadrant which provides answers to questions
asked in condition stub quadrant.
 Action Entry − It is in lower right quadrant which indicates the appropriate action
resulting from the answers to the conditions in the condition entry quadrant.
The entries in decision table are given by Decision Rules which define the relationships
between combinations of conditions and courses of action. In rules section,
 Y shows the existence of a condition.
 N represents the condition, which is not satisfied.
 A blank - against action states it is to be ignored.
 X (or a check mark will do) against action states it is to be carried out.
For example, refer the following table −

CONDITIONS Rule 1 Rule 2 Rule 3 Rule 4

Advance payment made Y N N N


Purchase amount = Rs 10,000/- - Y Y N

Regular Customer - Y N -

ACTIONS

Give 5% discount X X - -

Give no discount - - X X

Decision Trees
Decision trees are a method for defining complex relationships by describing decisions and
avoiding the problems in communication. A decision tree is a diagram that shows alternative
actions and conditions within horizontal tree framework. Thus, it depicts which conditions to
consider first, second, and so on.
Decision trees depict the relationship of each condition and their permissible actions. A square
node indicates an action and a circle indicates a condition. It forces analysts to consider the
sequence of decisions and identifies the actual decision that must be made.
The major limitation of a decision tree is
that it lacks information in its format to
describe what other combinations of
conditions you can take for testing. It is a
single representation of the relationships
between conditions and actions.
For example, refer the following decision
tree −

Data Flow Diagram

Data flow diagram is graphical representation of flow of data in an information system. It is


capable of depicting incoming data flow, outgoing data flow and stored data. The DFD does not
mention anything about how data flows through the system.
There is a prominent difference between DFD and Flowchart. The flowchart depicts flow of
control in program modules. DFDs depict flow of data in the system at various levels. DFD
does not contain any control or branch elements.

Types of DFD

Data Flow Diagrams are either Logical or Physical.

 Logical DFD - This type of DFD concentrates on the system process, and flow of data in
the system.For example in a Banking software system, how data is moved between
different entities.
 Physical DFD - This type of DFD shows how the data flow is actually implemented in
the system. It is more specific and close to the implementation.

The following table lists the points that differentiate a physical DFD from a logical DFD.

Physical DFD Logical DFD

It is implementation dependent. It shows which It is implementation independent. It focuses only on


functions are performed. the flow of data between processes.

It provides low level details of hardware, software, It explains events of systems and data required by
files, and people. each event.

It depicts how the current system operates and how It shows how business operates; not how the system
a system will be implemented. can be implemented.

DFD Components

DFD can represent Source, destination, storage and flow of data using the following set of
components -

 Entities - Entities are source and destination of information data. Entities are represented by a
rectangles with their respective names.
 Process - Activities and action taken on the data are represented by Circle or Round-edged rectangles.
 Data Storage - There are two variants of data storage - it can either be represented as a rectangle with
absence of both smaller sides or as an open-sided rectangle with only one side missing.
 Data Flow - Movement of data is shown by pointed arrows. Data movement is shown from the base
of arrow as its source towards head of the arrow as destination.

Levels of DFD
 Level 0 - Highest abstraction level DFD is Level 0 DFDs are also known as context
known as Level 0 DFD, which depicts the level DFDs.
entire information system as one diagram
concealing all the underlying details.
 Level 1 - The Level 0 DFD is broken down into
more specific, Level 1 DFD. Level 1 DFD
depicts basic modules in the system and flow of
data among various modules. Level 1 DFD also
mentions basic processes and sources of
information.

 Level 2 - At this level, DFD shows how data flows inside the modules mentioned in Level 1.
Higher level DFDs can be transformed into more specific lower level DFDs with deeper level of
understanding unless the desired level of specification is achieved.

Basic Elements of DFD

DFD is easy to understand and quite effective when the required design is not clear and the user
wants a notational language for communication. However, it requires a large number of
iterations for obtaining the most accurate and complete solution.
The following table shows the symbols used in designing a DFD and their significance −

Symbol Name Symbol Meaning

Square Source or Destination of Data

Arrow Data flow

Circle Process transforming data flow

Open Rectangle Data Store


Context Diagram
A context diagram helps in understanding the entire system by one DFD which gives the
overview of a system. It starts with mentioning major processes with little details and then goes
onto giving more details of the processes with the top-down approach.
The context diagram of mess management is shown below.

Entity-Relationship Model

Entity-Relationship model is a type of database model based on the notion of real world entities
and relationship among them. We can map real world scenario onto ER database model. ER
Model creates a set of entities with their attributes, a set of constraints and relation among them.
ER Model is best used for the conceptual design of database. ER Model can be represented as
follows :

 Entity - An entity in ER Model is a real world being, which has some properties
called attributes. Every attribute is defined by its corresponding set of values,
called domain.
For example, Consider a school database. Here, a student is an entity. Student has
various attributes like name, id, age and class etc.
The ER model defines the conceptual view of a database. It works around real-world entities
and the associations among them. At view level, the ER model is considered a good option for
designing databases.

Entity
An entity can be a real-world object, either animate or inanimate, that can be easily identifiable.
For example, in a school database, students, teachers, classes, and courses offered can be
considered as entities. All these entities have some attributes or properties that give them their
identity.
An entity set is a collection of similar types of entities. An entity set may contain entities with
attribute sharing similar values. For example, a Students set may contain all the students of a
school; likewise a Teachers set may contain all the teachers of a school from all faculties. Entity
sets need not be disjoint.

Attributes
Entities are represented by means of their properties, called attributes. All attributes have
values. For example, a student entity may have name, class, and age as attributes.
There exists a domain or range of values that can be assigned to attributes. For example, a
student's name cannot be a numeric value. It has to be alphabetic. A student's age cannot be
negative, etc.

Relationship
The association among entities is called a relationship. For example, an employee works_at a
department, a student enrolls in a course. Here, Works_at and Enrolls are called relationships.

Relationship Set
A set of relationships of similar type is called a relationship set. Like entities, a relationship too
can have attributes. These attributes are called descriptive attributes.

Entity Relationship Diagram

An Entity–relationship model (ER model) describes the structure of a database with the help of
a diagram, which is known as Entity Relationship Diagram (ER Diagram). An ER model is a
design or blueprint of a database that can later be implemented as a database. The main
components of E-R model are: entity set and relationship set.

A simple ER Diagram:

In the following diagram we have two Student and College is many to one as a
entities Student and College and their college can have many students however a
relationship. The relationship between student cannot study in multiple colleges at
the same time. Student entity has attributes
such as Stu_Id, Stu_Name & Stu_Addr and
College entity has attributes such as Col_ID
& Col_Name.

Here are the geometric shapes and their meaning in an E-R Diagram. We will discuss these terms
in detail in the next section(Components of a ER Diagram) of this guide so don’t worry too much
about these terms now, just go through them once.

Rectangle : Represents Entity sets.


Ellipses : Attributes
Diamonds : Relationship Set
Lines : They link attributes to Entity Sets and Entity sets to Relationship Set
Double Ellipses: Multivalued Attributes
Dashed Ellipses: Derived Attributes
Double Rectangles: Weak Entity Sets
Double Lines : Total participation of an entity in a relationship set

Components of a ER Diagram

As shown in the diagram, an ER diagram


has three main components:
1. Entity
2. Attribute
3. Relationship

1. Entity
An entity is an object or component of data. An entity is represented as rectangle in an ER
diagram.
For example: In the following ER diagram
we have two entities Student and College
and these two entities have many to one
relationship as many students study in a
single college. We will read more about
relationships later, for now focus on entities.

Weak Entity:
An entity that cannot be uniquely identified
by its own attributes and relies on the
relationship with other entity is called weak
entity. The weak entity is represented by a
double rectangle. For example – a bank
account cannot be uniquely identified
without knowing the bank to which the
account belongs, so bank account is a weak
entity.
2. Attribute
An attribute describes the property of an entity. An attribute is represented as Oval in an ER
diagram. There are four types of attributes:
1. Key attribute
2. Composite attribute
3. Multivalued attribute
4. Derived attribute
1. Key attribute:
A key attribute can uniquely identify an
entity from an entity set. For example,
student roll number can uniquely identify a
student from a set of students. Key attribute
is represented by oval same as other
attributes however the text of key attribute
is underlined.

2. Composite attribute:
An attribute that is a combination of other
attributes is known as composite attribute.
For example, In student entity, the student
address is a composite attribute as an
address is composed of other attributes such
as pin code, state, country.
3. Multivalued attribute:

An attribute that can hold multiple values is known as multivalued attribute. It is represented
with double ovals in an ER Diagram. For example – A person can have more than one phone
numbers so the phone number attribute is multivalued.

4. Derived attribute:

A derived attribute is one whose value is E-R diagram with multivalued and
dynamic and derived from another attribute. derived attributes:
It is represented by dashed oval in an ER
Diagram. For example – Person age is a
derived attribute as it changes over time and
can be derived from another attribute (Date
of birth).

3. Relationship
A relationship is represented by diamond shape in ER diagram, it shows the relationship among
entities. There are four types of relationships:
1. One to One 3. Many to One
2. One to Many 4. Many to Many

1. One to One Relationship


When a single instance of an entity is
associated with a single instance of another
entity then it is called one to one
relationship. For example, a person has only
one passport and a passport is given to one
person.

2. One to Many Relationship

When a single instance of an entity is


associated with more than one instances of
another entity then it is called one to many
relationship. For example – a customer can
place many orders but a order cannot be
placed by many customers.
3. Many to One Relationship
When more than one instances of an entity is
associated with a single instance of another
entity then it is called many to one
relationship. For example – many students
can study in a single college but a student
cannot study in many colleges at the same
time.

4. Many to Many Relationship

When more than one instances of an entity is


associated with more than one instances of
another entity then it is called many to many
relationship. For example, a can be assigned
to many projects and a project can be
assigned to many students.

Total Participation of an Entity set

A Total participation of an entity set


represents that each entity in entity set must
have at least one relationship in a
relationship set. For example: In the below
diagram each college must have at-least one
associated Student.

Generalization is a process in which the common attributes of more than one entities form a
new entity. This newly formed entity is called generalized entity.
Generalization Example
Lets say we have two entities Student and The ER diagram before generalization
Teacher. looks like this:
Attributes of Entity Student are: Name,
Address & Grade
Attributes of Entity Teacher are: Name,
Address & Salary

These two entities have two common attributes: Name and Address, we can make a generalized
entity with these common attributes. Lets have a look at the ER model after generalization.
OOAD - Object Oriented Analysis & Design
A Brief History
The object-oriented paradigm took its shape from the initial concept of a new programming
approach, while the interest in design and analysis methods came much later.
 The first object–oriented language was Simula (Simulation of real systems) that was developed in
1960 by researchers at the Norwegian Computing Center.
 In 1970, Alan Kay and his research group at Xerox PARK created a personal computer named
Dynabook and the first pure object-oriented programming language (OOPL) - Smalltalk, for
programming the Dynabook.
 In the 1980s, Grady Booch published a paper titled Object Oriented Design that mainly presented a
design for the programming language, Ada. In the ensuing editions, he extended his ideas to a
complete object–oriented design method.
 In the 1990s, Coad incorporated behavioral ideas to object-oriented methods.
The other significant innovations were Object Modelling Techniques (OMT) by James
Rumbaugh and Object-Oriented Software Engineering (OOSE) by Ivar Jacobson.

Object-Oriented Analysis
Object–Oriented Analysis (OOA) is the procedure of identifying software engineering
requirements and developing software specifications in terms of a software system’s object
model, which comprises of interacting objects.
The main difference between object-oriented analysis and other forms of analysis is that in
object-oriented approach, requirements are organized around objects, which integrate both data
and functions. They are modelled after real-world objects that the system interacts with. In
traditional analysis methodologies, the two aspects - functions and data - are considered
separately.
Grady Booch has defined OOA as, “Object-oriented analysis is a method of analysis that
examines requirements from the perspective of the classes and objects found in the vocabulary
of the problem domain”.
The primary tasks in object-oriented analysis (OOA) are −

 Identifying objects
 Organizing the objects by creating object model diagram
 Defining the internals of the objects, or object attributes
 Defining the behavior of the objects, i.e., object actions
 Describing how the objects interact
The common models used in OOA are use cases and object models.

Object-Oriented Design
Object–Oriented Design (OOD) involves implementation of the conceptual model produced
during object-oriented analysis. In OOD, concepts in the analysis model, which are
technology−independent, are mapped onto implementing classes, constraints are identified and
interfaces are designed, resulting in a model for the solution domain, i.e., a detailed description
of how the system is to be built on concrete technologies.
The implementation details generally include −

 Restructuring the class data (if necessary),


 Implementation of methods, i.e., internal data structures and algorithms,
 Implementation of control, and
 Implementation of associations.
Grady Booch has defined object-oriented design as “a method of design encompassing the
process of object-oriented decomposition and a notation for depicting both logical and physical
as well as static and dynamic models of the system under design”.

Object-Oriented Programming
Object-oriented programming (OOP) is a programming paradigm based upon objects (having
both data and methods) that aims to incorporate the advantages of modularity and reusability.
Objects, which are usually instances of classes, are used to interact with one another to design
applications and computer programs.
The important features of object–oriented programming are −

 Bottom–up approach in program design


 Programs organized around objects, grouped in classes
 Focus on data with methods to operate upon object’s data
 Interaction between objects through functions
 Reusability of design through creation of new classes by adding features to existing classes
Some examples of object-oriented programming languages are C++, Java, Smalltalk, Delphi,
C#, Perl, Python, Ruby, and PHP.
Grady Booch has defined object–oriented programming as “a method of implementation in
which programs are organized as cooperative collections of objects, each of which represents
an instance of some class, and whose classes are all members of a hierarchy of classes united
via inheritance relationships”.
The object model visualizes the elements in a software application in terms of objects. In this
chapter, we will look into the basic concepts and terminologies of object–oriented systems.

Objects and Classes


The concepts of objects and classes are intrinsically linked with each other and form the
foundation of object–oriented paradigm.

Object
An object is a real-world element in an object–oriented environment that may have a physical or
a conceptual existence. Each object has −
 Identity that distinguishes it from other objects in the system.
 State that determines the characteristic properties of an object as well as the values of the properties
that the object holds.
 Behavior that represents externally visible activities performed by an object in terms of changes in its
state.
Objects can be modelled according to the needs of the application. An object may have a
physical existence, like a customer, a car, etc.; or an intangible conceptual existence, like a
project, a process, etc.

Class
A class represents a collection of objects having same characteristic properties that exhibit
common behavior. It gives the blueprint or description of the objects that can be created from it.
Creation of an object as a member of a class is called instantiation. Thus, object is an instance of
a class.
The constituents of a class are −
 A set of attributes for the objects that are to be instantiated from the class. Generally, different objects
of a class have some difference in the values of the attributes. Attributes are often referred as class
data.
 A set of operations that portray the behavior of the objects of the class. Operations are also referred as
functions or methods.

UML
UML (Unified Modeling Language) is a standard language for specifying, visualizing, constructing,
and documenting the artifacts of software systems. UML was created by the Object Management
Group (OMG) and UML 1.0 specification draft was proposed to the OMG in January 1997. It was
initially started to capture the behavior of complex software and non-software system and now it has
become an OMG standard.

UML is a standard language for specifying, visualizing, constructing, and documenting the
artifacts of software systems.
UML was created by the Object Management Group (OMG) and UML 1.0 specification draft
was proposed to the OMG in January 1997.
OMG is continuously making efforts to create a truly industry standard.
 UML stands for Unified Modeling Language.
 UML is different from the other common programming languages such as C++, Java, COBOL, etc.
 UML is a pictorial language used to make software blueprints.
 UML can be described as a general purpose visual modeling language to visualize, specify, construct,
and document software system.
 Although UML is generally used to model software systems, it is not limited within this boundary. It
is also used to model non-software systems as well. For example, the process flow in a
manufacturing unit, etc.
UML is not a programming language but tools can be used to generate code in various
languages using UML diagrams. UML has a direct relation with object oriented analysis and
design. After some standardization, UML has become an OMG standard.

Goals of UML
A picture is worth a thousand words, this idiom absolutely fits describing UML. Object-
oriented concepts were introduced much earlier than UML. At that point of time, there were no
standard methodologies to organize and consolidate the object-oriented development. It was
then that UML came into picture.
There are a number of goals for developing UML but the most important is to define some
general purpose modeling language, which all modelers can use and it also needs to be made
simple to understand and use.
UML diagrams are not only made for developers but also for business users, common people,
and anybody interested to understand the system. The system can be a software or non-software
system. Thus it must be clear that UML is not a development method rather it accompanies with
processes to make it a successful system.
In conclusion, the goal of UML can be defined as a simple modeling mechanism to model all
possible practical systems in today’s complex environment.

A Conceptual Model of UML


To understand the conceptual model of UML, first we need to clarify what is a conceptual
model? and why a conceptual model is required?
 A conceptual model can be defined as a model which is made of concepts and their relationships.
 A conceptual model is the first step before drawing a UML diagram. It helps to understand the
entities in the real world and how they interact with each other.
As UML describes the real-time systems, it is very important to make a conceptual model and
then proceed gradually. The conceptual model of UML can be mastered by learning the
following three major elements −
 UML building blocks
 Rules to connect the building blocks
 Common mechanisms of UML

Role of UML in OO Design


UML is a modeling language used to model software and non-software systems. Although
UML is used for non-software systems, the emphasis is on modeling OO software applications.
Most of the UML diagrams discussed so far are used to model different aspects such as static,
dynamic, etc. Now whatever be the aspect, the artifacts are nothing but objects.
If we look into class diagram, object diagram, collaboration diagram, interaction diagrams all
would basically be designed based on the objects.
Hence, the relation between OO design and UML is very important to understand. The OO
design is transformed into UML diagrams according to the requirement. Before understanding
the UML in detail, the OO concept should be learned properly. Once the OO analysis and
design is done, the next step is very easy. The input from OO analysis and design is the input to
UML diagrams.

UML Diagrams
UML diagrams are the ultimate output of the entire discussion. All the elements, relationships
are used to make a complete UML diagram and the diagram represents a system.
The visual effect of the UML diagram is the most important part of the entire process. All the
other elements are used to make it complete.
UML includes the following nine diagrams, the details of which are described in the subsequent
chapters.

 Class diagram
 Object diagram
 Use case diagram
 Sequence diagram
 Collaboration diagram
 Activity diagram
 State chart diagram
 Deployment diagram
 Component diagram
Unit III: Database Management Systems

Syllabus: DBMS – types and evolution, RDBMS, OODBMS, RODBMS, Data warehousing,
Data Mart, Data mining

Database - definition
A database is an organized collection of structured information, or data, typically stored
electronically in a computer system. A database is usually controlled by a database management
system (DBMS).
What is Database?
A database is a systematic collection of data. They support electronic storage and manipulation
of data. Databases make data management easy.
Together, the data and the DBMS, along with the applications that are associated with them, are
referred to as a database system, often shortened to just database.
Let us discuss a database example: Facebook. It needs to store, manipulate, and present data
related to members, their friends, member activities, messages, advertisements, and a lot more.
We can provide a countless number of examples for the usage of databases.
Types of Databases
Here are some popular types of databases.
1. Distributed databases:
A distributed database is a type of database that has contributions from the common database and
information captured by local computers. In this type of database system, the data is not in one
place and is distributed at various organizations.
2. Relational databases:
This type of database defines database relationships in the form of tables. It is also called
Relational DBMS, which is the most popular DBMS type in the market. Database example of the
RDBMS system include MySQL, Oracle, and Microsoft SQL Server database.
3. Object-oriented databases:
This type of computers database supports the storage of all data types. The data is stored in the
form of objects. The objects to be held in the database have attributes and methods that define
what to do with the data. PostgreSQL is an example of an object-oriented relational DBMS.
4. Centralized database:
It is a centralized location, and users from different backgrounds can access this data. This type
of computers databases store application procedures that help users access the data even from a
remote location.
5. Data warehouses:
Data Warehouse is to facilitate a single version of truth for a company for decision making and
forecasting. A Data warehouse is an information system that contains historical and commutative
data from single or multiple sources. Data Warehouse concept simplifies the reporting and
analysis process of the organization.
6. NoSQL databases:
NoSQL database is used for large sets of distributed data. There are a few big data performance
problems that are effectively handled by relational databases. This type of computers database is
very efficient in analyzing large-size unstructured data.
7. Graph databases:
A graph-oriented database uses graph theory to store, map, and query relationships. These kinds
of computers databases are mostly used for analyzing interconnections. For example, an
organization can use a graph database to mine data about customers from social media.
8. OLTP databases:
OLTP another database type which able to perform fast query processing and maintaining data
integrity in multi-access environments.
9. Personal database:
A personal database is used to store data stored on personal computers that are smaller and easily
manageable. The data is mostly used by the same department of the company and is accessed by
a small group of people.
10. Hierarchical:
This type of DBMS employs the “parent-child” relationship of storing data. Its structure is like a
tree with nodes representing records and branches representing fields. The windows registry used
in Windows XP is a hierarchical database example.
11. Network DBMS:
This type of DBMS supports many-to-many relations. It usually results in complex database
structures. RDM Server is an example of database management system that implements the
network model.
Some of the latest databases include
12. Open-source databases:
This kind of database stored information related to operations. It is mainly used in the field of
marketing, employee relations, customer service, of databases.
13. Cloud databases:
A cloud database is a database which is optimized or built for such a virtualized environment.
There are so many advantages of a cloud database, some of which can pay for storage capacity
and bandwidth. It also offers scalability on-demand, along with high availability.
14. Self-driving databases:
The newest and most groundbreaking type of database, self-driving databases (also known as
autonomous databases) are cloud-based and use machine learning to automate database
tuning, security, backups, updates, and other routine management tasks traditionally
performed by database administrators.
15. Multimodal database:
The multimodal database is a type of data processing platform that supports multiple data models
that define how the certain knowledge and information in a database should be organized and
arranged.
16. Document/JSON database:
In a document-oriented database, the data is kept in document collections, usually using the
XML, JSON, BSON formats. One record can store as much data as you want, in any data type
(or types) you prefer.

Database Components

There are five main components of a database:


Hardware:
The hardware consists of physical, electronic devices like computers, I/O devices, storage
devices, etc. This offers the interface between computers and real-world systems.
Software:
This is a set of programs used to manage and control the overall database. This includes the
database software itself, the Operating System, the network software used to share the data
among users, and the application programs for accessing data in the database.
Data:
Data is a raw and unorganized fact that is required to be processed to make it meaningful. Data
can be simple at the same time unorganized unless it is organized. Generally, data comprises
facts, observations, perceptions, numbers, characters, symbols, images, etc.
Procedure:
Procedure are a set of instructions and rules that help you to use the DBMS. It is designing and
running the database using documented methods, which allows you to guide the users who
operate and manage it.
Database Access Language:
Database Access language is used to access the data to and from the database, enter new data,
update already existing data, or retrieve required data from DBMS. The user writes some specific
commands in a database access language and submits these to the database.

DBMS
DBMS stands for Database Management System. We can break it like this DBMS =
Database + Management System.
A database management system stores data in such a way that it becomes easier to
retrieve, manipulate, and produce information. DBMS is a collection of inter-related data
and set of programs to store & access those data in an easy and effective manner.

What is database software?


Database software is used to create, edit, and maintain database files and records, enabling easier
file and record creation, data entry, data editing, updating, and reporting. The software also
handles data storage, backup and reporting, multi-access control, and security. Database software
is sometimes also referred to as a “database management system” (DBMS).

Database software makes data management simpler by enabling users to store data in a
structured form and then access it. It typically has a graphical interface to help create and
manage the data and, in some cases, users can construct their own databases by using database
software.

What is a Database Management System (DBMS)?


Database Management System (DBMS) is a collection of programs that enable its users to
access databases, manipulate data, report, and represent data. It also helps to control access to the
database. Database Management Systems are not a new concept and, as such, had been first
implemented in the 1960s.
Evolution of DBMS
Charles Bachman’s Integrated Data Store (IDS) is said to be the first DBMS in history. With
time database, technologies evolved a lot, while usage and expected functionalities of databases
increased immensely.
History of Database Management System
Here, are the important landmarks from the history:
 1960 – Charles Bachman designed first DBMS system.
 1970 – Codd introduced IBM’S Information Management System (IMS).
 1976 – Peter Chen coined and defined the Entity-relationship model also know as the ER
model.
 1980 – Relational model becomes a widely accepted database component.
 1985 – Object-oriented DBMS develops.
 1990 – Incorporation of object-orientation in relational DBMS.
 1991 – Microsoft ships MS access, a personal DBMS and that displaces all other personal
DBMS products.
 1995 – First Internet database applications.
 1997 – XML applied to database processing. Many vendors begin to integrate XML into
DBMS products.
Advantages of DBMS
 DBMS offers a variety of techniques to store & retrieve data.
 DBMS serves as an efficient handler to balance the needs of multiple applications using
the same data.
 Uniform administration procedures for data.
 Application programmers never exposed to details of data representation and storage.
 A DBMS uses various powerful functions to store and retrieve data efficiently.
 Offers Data Integrity and Security.
 The DBMS implies integrity constraints to get a high level of protection against
prohibited access to data.
 A DBMS schedules concurrent access to the data in such a manner that only one user can
access the same data at a time.
 Reduced Application Development Time.
Disadvantage of DBMS
DBMS may offer plenty of advantages but, it has certain flaws-
 Cost of Hardware and Software of a DBMS is quite high which increases the budget of
your organization.
 Most database management systems are often complex systems, so the training for users
to use the DBMS is required.
 In some organizations, all data is integrated into a single database which can be damaged
because of electric failure or database is corrupted on the storage media.
 Use of the same program at a time by many users sometimes lead to the loss of some
data.
 DBMS can’t perform sophisticated calculations.
A database typically requires a comprehensive database software program known as a database
management system (DBMS). A DBMS serves as an interface between the database and its end
users or programs, allowing users to retrieve, update, and manage how the information is
organized and optimized. A DBMS also facilitates oversight and control of databases, enabling a
variety of administrative operations such as performance monitoring, tuning, and backup and
recovery.

Examples of popular database software or DBMSs include MySQL, Microsoft Access,


Microsoft SQL Server, FileMaker Pro, Oracle Database, and dBASE.
Need of DBMS
Database systems are basically developed for large amount of data. When dealing with
huge amount of data, there are two things that require optimization:
(i) Storage of data and (ii) retrieval of data.
Storage: According to the principles of database systems, the data is stored in such a way
that it acquires lot less space as the redundant data (duplicate data) has been removed
before storage. Let’s take a layman example to understand this:
In a banking system, suppose a customer is having two accounts, one is saving account
and another is salary account. Let’s say bank stores saving account data at one place
(these places are called tables we will learn them later) and salary account data at another
place, in that case if the customer information such as customer name, address etc. are
stored at both places then this is just a wastage of storage (redundancy/ duplication of
data), to organize the data in a better way the information should be stored at one place
and both the accounts should be linked to that information somehow. The same thing we
achieve in DBMS.
Fast Retrieval of data: Along with storing the data in an optimized and systematic
manner, it is also important that we retrieve the data quickly when needed. Database
systems ensure that the data is retrieved as quickly as possible.
Purpose of Database Systems
The main purpose of database systems is to manage the data. Consider a university that
keeps the data of students, teachers, courses, books etc. To manage this data we need to
store this data somewhere where we can add new data, delete unused data, update
outdated data, retrieve data, to perform these operations on data we need a Database
management system that allows us to store the data in such a way so that all these
operations can be performed on the data efficiently.
Applications where we use Database Management Systems are:

 Telecom: There is a database to keeps track of the information regarding calls


made, network usage, customer details etc. Without the database systems it is hard
to maintain that huge amount of data that keeps updating every millisecond.
 Industry: Where it is a manufacturing unit, warehouse or distribution centre, each
one needs a database to keep the records of ins and outs. For example distribution
centre should keep a track of the product units that supplied into the centre as well
as the products that got delivered out from the distribution centre on each day; this
is where DBMS comes into picture.
 Banking System: For storing customer info, tracking day to day credit and debit
transactions, generating bank statements etc. All this work has been done with the
help of Database management systems.
 Sales: To store customer information, production information and invoice details.
 Airlines: To travel though airlines, we make early reservations, this reservation
information along with flight schedule is stored in database.
 Education sector: Database systems are frequently used in schools and colleges
to store and retrieve the data regarding student details, staff details, course details,
exam details, payroll data, attendance details, fees details etc. There is a hell lot
amount of inter-related data that needs to be stored and retrieved in an efficient
manner.
 Online shopping: You must be aware of the online shopping websites such as
Amazon, Flipkart etc. These sites store the product information, your addresses and
preferences, credit details and provide you the relevant list of products based on
your query. All this involves a Database management system.
Drawbacks of File system
 Data redundancy: Data redundancy refers to the duplication of data, lets say we are
managing the data of a college where a student is enrolled for two courses, the same
student details in such case will be stored twice, which will take more storage than
needed. Data redundancy often leads to higher storage costs and poor access time.
 Data inconsistency: Data redundancy leads to data inconsistency, lets take the same
example that we have taken above, a student is enrolled for two courses and we
have student address stored twice, now lets say student requests to change his
address, if the address is changed at one place and not on all the records then this
can lead to data inconsistency.
 Data Isolation: Because data are scattered in various files, and files may be in
different formats, writing new application programs to retrieve the appropriate data
is difficult.
 Dependency on application programs: Changing files would lead to change in
application programs.
 Atomicity issues: Atomicity of a transaction refers to “All or nothing”, which means
either all the operations in a transaction executes or none.
 Data Security: Data should be secured from unauthorised access, for example a
student in a college should not be able to see the payroll details of the teachers, such
kind of security constraints are difficult to apply in file processing systems.
Advantage of DBMS over file system
There are several advantages of Database management system over file system. Few of
them are as follows:

 No redundant data: Redundancy removed by data normalization. No data


duplication saves storage and improves access time.
 Data Consistency and Integrity: As we discussed earlier the root cause of data
inconsistency is data redundancy, since data normalization takes care of the data
redundancy, data inconsistency also been taken care of as part of it
 Data Security: It is easier to apply access constraints in database systems so that
only authorized user is able to access the data. Each user has a different set of
access thus data is secured from the issues such as identity theft, data leaks and
misuse of data.
 Privacy: Limited access means privacy of data.
 Easy access to data – Database systems manages data in such a way so that the data
is easily accessible with fast response times.
 Easy recovery: Since database systems keeps the backup of data, it is easier to do a
full recovery of data in case of a failure.
 Flexible: Database systems are more flexible than file processing systems.
DBMS Architecture
The architecture of DBMS depends on the computer system on which it runs. For example, in a
client-server DBMS architecture, the database systems at server machine can run several
requests made by client machine. We will understand this communication with the help of
diagrams.
Types of DBMS Architecture
There are three
types of DBMS architecture:
1. Single tier architecture
2. Two tier architecture
3. Three tier architecture
1. Single tier architecture
In this type of architecture, the database is readily available on the client machine, any request
made by client doesn’t require a network connection to perform the action on the database.
For example, lets say you want to fetch the records of employee from the database and the
database is available on your computer system, so the request to fetch employee details will be
done by your computer and the records will be fetched from the database by your computer as
well. This type of system is generally referred as local database system.
2. Two tier architecture

In two-tier architecture, the Database system is present at the server machine and the DBMS
application is present at the client machine, these two machines are connected with each other
through a reliable network as shown in the above diagram.
Whenever client machine makes a request to access the database present at server using a query
language like sql, the server perform the request on the database and returns the result back to the
client. The application connection interface such as JDBC, ODBC are used for the interaction
between server and client.
3. Three tier architecture

In three-tier architecture, another layer is present between the client machine and server machine.
In this architecture, the client application doesn’t communicate directly with the database
systems present at the server machine, rather the client application communicates with server
application and the server application internally communicates with the database system present
at the server
.
DBMS – Three Level Architecture
DBMS Three Level Architecture Diagram

This architecture has three levels:


1. External level
2. Conceptual level
3. Internal level
1. External level
It is also called view level. The reason this level is called “view” is because several users can
view their desired data from this level which is internally fetched from database with the help of
conceptual and internal level mapping.
The user doesn’t need to know the database schema details such as data structure, table definition
etc. user is only concerned about data which is what returned back to the view level after it has
been fetched from database (present at the internal level).
External level is the “top level” of the Three Level DBMS Architecture.
2. Conceptual level
It is also called logical level. The whole design of the database such as relationship among data,
schema of data etc. are described in this level.
Database constraints and security are also implemented in this level of architecture. This level is
maintained by DBA (database administrator).
3. Internal level
This level is also known as physical level. This level describes how the data is actually stored in
the storage devices. This level is also responsible for allocating space to the data. This is the
lowest level of the architecture.
View of Data in DBMS
Abstraction is one of the main features of database systems. Hiding irrelevant details from user
and providing abstract view of data to users, helps in easy and efficient user-
database interaction.
To fully understand the view of data, you must have a basic knowledge of data abstraction and
instance & schema.
Data Abstraction in DBMS
Database systems are made-up of complex data structures. To ease the user interaction with
database, the developers hide internal irrelevant details from users. This process of hiding
irrelevant details from user is called data abstraction.

We have three levels of abstraction:


Physical level: This is the lowest level of data abstraction. It describes how data is actually
stored in database. You can get the complex data structure details at this level.
Logical level: This is the middle level of 3-level data abstraction architecture. It describes what
data is stored in database.
View level: Highest level of data abstraction. This level describes the user interaction with
database system.
Example: Let’s say we are storing customer information in a customer table. At physical
level these records can be described as blocks of storage (bytes, gigabytes, terabytes etc.) in
memory. These details are often hidden from the programmers.
At the logical level these records can be described as fields and attributes along with their data
types, their relationship among each other can be logically implemented. The programmers
generally work at this level because they are aware of such things about database systems.
At view level, user just interact with system with the help of GUI and enter the details at the
screen, they are not aware of how the data is stored and what data is stored; such details are
hidden from them.
DBMS Schema
Definition of schema: Design of a database is called the schema. Schema is of three types:
Physical schema, logical schema and view schema.
For example: In the following diagram, we have a schema that shows the relationship between
three tables: Course, Student and Section. The diagram only shows the design of the database, it
doesn’t show the data present in those tables. Schema is only a structural view(design) of a
database as shown in the diagram below.

The design of a database at physical level is called physical schema, how the data stored in
blocks of storage is described at this level.
Design of database at logical level is called logical schema, programmers and database
administrators work at this level, at this level data can be described as certain types of data
records gets stored in data structures, however the internal details such as implementation of data
structure is hidden at this level (available at physical level).
Design of database at view level is called view schema. This generally describes end user
interaction with database systems.
DBMS Instance
Definition of instance: The data stored in database at a particular moment of time is called
instance of database. Database schema defines the variable declarations in tables that belong to a
particular database; the value of these variables at a moment of time is called the instance of that
database.
For example, lets say we have a single table student in the database, today the table has 100
records, so today the instance of the database has 100 records. Lets say we are going to add
another 100 records in this table by tomorrow so the instance of database tomorrow will have
200 records in table. In short, at a particular moment the data stored in database is called the
instance, that changes over time when we add or delete data from the database.
DBMS languages
Database languages are used to read, update and store data in a database. There are several such
languages that can be used for this purpose; one of them is SQL (Structured Query Language).
Types of DBMS languages:

Data Definition Language (DDL)


DDL is used for specifying the database schema. It is used for creating tables, schema, indexes,
constraints etc. in database. Lets see the operations that we can perform on database using DDL:

 To create the database instance – CREATE


 To alter the structure of database – ALTER
 To drop database instances – DROP
 To delete tables in a database instance – TRUNCATE
 To rename database instances – RENAME
 To drop objects from database such as tables – DROP
 To Comment – Comment
All of these commands either defines or update the database schema that’s why they come under
Data Definition language.
Data Manipulation Language (DML)
DML is used for accessing and manipulating data in a database. The following operations on
database comes under DML:

 To read records from table(s) – SELECT


 To insert record(s) into the table(s) – INSERT
 Update the data in table(s) – UPDATE
 Delete all the records from the table – DELETE
Data Control language (DCL)
DCL is used for granting and revoking user access on a database –

 To grant access to user – GRANT


 To revoke access from user – REVOKE
In practical data definition language, data manipulation language and data control languages are
not separate language, rather they are the parts of a single database language such as SQL.
Transaction Control Language(TCL)
The changes in the database that we made using DML commands are either performed or
rollbacked using TCL.

 To persist the changes made by DML commands in database – COMMIT


 To rollback the changes made to the database – ROLLBACK
Data models in DBMS
Data Model is a logical structure of Database. It describes the design of database to reflect
entities, attributes, relationship among data, constrains etc.
Types of Data Models
There are several types of data models in DBMS.
I. Object based logical Models – Describe data at the conceptual and view levels.
1. E-R Model 2. Object oriented Model
(OODBMS)
II. Record based logical Models – Like Object based model, they also describe data at the
conceptual and view levels. These models specify logical structure of database with
records, fields and attributes.
1. Relational Model (RDBMS) 2. Hierarchical Model (HDBMS)
3. Network Model- Network Model is same as hierarchical model except that it has graph-
like structure rather than a tree-based structure. Unlike hierarchical model, this model
allows each record to have more than one parent record.
4. Relation- Object DBMS (RODBMS)
Entity Relationship Diagram – ER Diagram in DBMS
An Entity–relationship model (ER model) describes the structure of a database with the help of
a diagram, which is known as Entity Relationship Diagram (ER Diagram). An ER model is a
design or blueprint of a database that can later be implemented as a database. The main
components of E-R model are: entity set and relationship set.
What is an Entity Relationship Diagram (ER Diagram)?
An ER diagram shows the relationship among entity sets. An entity set is a group of similar
entities and these entities can have attributes. In terms of DBMS, an entity is a table or attribute
of a table in database, so by showing relationship among tables and their attributes, ER diagram
shows the complete logical structure of a database.
We have a detailed notes on this in the II Unit notes (pls refer II unit notes)
Relational model in DBMS (RDBMS)
In relational model, the data and relationships are represented by collection of inter-related
tables. Each table is a group of column and rows, where column represents attribute of an entity
and rows represents records.
Sample relationship Model: Student table with 3 columns and four records.
Table: Student

Stu_Id Stu_Name Stu_Age

111 Ashish 23

123 Saurav 22

169 Lester 24

234 Lou 26

Table: Course

Stu_Id Course_Id Course_Name


111 C01 Science
111 C02 DBMS
169 C22 Java
169 C39 Computer
Here Stu_Id, Stu_Name & Stu_Age are attributes of table Student and Stu_Id, Course_Id &
Course_Name are attributes of table Course. TheNetworks
rows with values are the records (commonly
known as tuples).
Hierarchical model in DBMS (HDBMS)
In hierarchical model, data is organized into a tree like structure with each record is having one
parent record and many children. The main drawback of this model is that, it can have only one
to many relationships between nodes.
Note: Hierarchical models are rarely used now.
Sample Hierarchical Model Diagram:
Lets say we have few students and few courses and a course can be assigned to a single student
only, however a student take any number of courses so this relationship becomes one to many.

Example of hierarchical data represented as


relational tables: The above hierarchical model can be represented as relational tables like this:
Student Table

Stu_Id Stu_Name Stu_Age

123 Steve 29

367 Chaitanya 27

234 Ajeet 28
Course Table:

Course_Id Course_Name Stu_Id

C01 Cobol 123

C21 Java 367

C22 Perl 367

C33 JQuery 234


Types of Data Model

Following are the types of Data Model,


1. Hierarchical Model
2. Relational Model
3. Network Database Model
4. Entity Relationship Model
5. Object Model
1. Hierarchical Model
 Hierarchical model was developed by IBM and North American Rockwell known as
Information Management System.
 It represents the data in a hierarchical tree structure.
 This model is the first DBMS model.
 In this model, the data is sorted hierarchically.
 It uses pointer to navigate between the stored data.

2. Relational Model

 Relational model is based on first-order predicate logic.


 This model was first proposed by E. F. Codd.
 It represents data as relations or tables.
 Relational database simplifies the database structure by making use of tables and columns.
3. Network Database Model

 Network Database Model is same like Hierarchical Model, but the only difference is that it
allows a record to have more than one parent.
 In this model, there is no need of parent to child association like the hierarchical model.
 It replaces the hierarchical tree with a graph.
 It represents the data as record types and one-to-many relationship.
 This model is easy to design and understand.

4. Entity Relationship Model

 Entity Relationship Model is a high-level data model.


 It was developed by Chen in 1976.
 This model is useful in developing a conceptual design for the database.
 It is very simple and easy to design logical view of data.
 The developer can easily understand the system by looking at an ER model constructed .
In this diagram,

 Rectangle represents the entities. Eg. Doctor and Patient.


 Ellipse represents the attributes. Eg. DocId, Dname, PId, Pname. Attribute describes each
entity becomes a major part of the data stored in the database.
 Diamond represents the relationship in ER diagrams. Eg. Doctor diagnoses the Patient.

5. Object Model

 Object model stores the data in the form of objects, classes and inheritance.
 This model handles more complex applications, such as Geographic Information System
(GIS), scientific experiments, engineering design and manufacturing.
 It is used in File Management System.
 It represents real world objects, attributes and behaviors.
 It provides a clear modular structure.
 It is easy to maintain and modify the existing code.
RDBMS Concepts
RDBMS stands for relational database management system. A relational model can be
represented as a table of rows and columns. A relational database has following major
components:

1. Table 5. Instance
2. Record or Tuple 6. Schema
3. Field or Column name or Attribute 7. Keys
4. Domain

1. Table
A table is a collection of data represented in rows and columns. Each table has a name in
database. For example, the following table “STUDENT” stores the information of students in
database.
Table: STUDENT

Student_Id Student_Name Student_Addr Student_Age

101 Chaitanya Dayal Bagh, Agra 27


102 Ajeet Delhi 26
103 Rahul Gurgaon 24
104 Shubham Chennai 25
2. Record or Tuple
Each row of a table is known as record. It is also known as tuple. For example, the following row
is a record that we have taken from the above table.

102 Ajeet Delhi 26


3. Field or Column name or Attribute
The above table “STUDENT” has four fields (or attributes): Student_Id, Student_Name,
Student_Addr & Student_Age.
4. Domain
A domain is a set of permitted values for an attribute in table. For example, a domain of month-
of-year can accept January, February,…December as values, a domain of dates can accept all
possible valid dates etc. We specify domain of attribute while creating a table.
An attribute cannot accept values that are outside of their domains. For example, In the above
table “STUDENT”, the Student_Id field has integer domain so that field cannot accept values
that are not integers for example, Student_Id cannot has values like, “First”, 10.11 etc.
5. Instance
The data stored in database at a particular moment of time is called instance of database. The
value of these variables at a moment of time is called the instance of that database.
6. Schema
Database schema defines the variable declarations in tables that belong to a particular database.
The design of a database at physical level is called physical schema, how the data stored in
blocks of storage is described at this level Design of database at logical level is called logical
schema. Design of database at view level is called view schema. This generally describes end
user interaction with database systems.
7. Keys
Key plays an important role in relational database; it is used for identifying unique rows from
table. It also establishes relationship among tables.
Types of keys in DBMS
Primary Key – A primary is a column or set of columns in a table that uniquely identifies tuples
(rows) in that table.
Super Key – A super key is a set of one of more columns (attributes) to uniquely identify rows
in a table.
Candidate Key – A super key with no redundant attribute is known as candidate key

Alternate Key – Out of all candidate keys, only one gets selected as primary key, remaining
keys are known as alternate or secondary keys.
Composite Key – A key that consists of more than one attribute to uniquely identify rows (also
known as records & tuples) in a table is called composite key.
Foreign Key – Foreign keys are the columns of a table that points to the primary key of another
table. They act as a cross-reference between tables.

Difference between RDBMS and OODBMS


RDBMS and OODBMS are database management systems. RDBMS uses tables to represent
data and their relationships whereas OODBMS represents data in form of objects similar to
Object Oriented Programming.
Following are the important differences between RDBMS and OODBMS.

Sr. Key RDBMS OODBMS


No.

1 Definition RDBMS stands for Relational DataBase OODBMS stands for Object Oriented DataBase
Management System. Management System.

2 Data Data is stored as entities defined in Data is stored as objects.


Management tabular format.

3 Data RDBMS handles simple data. OODBMS handles large and complex data.
Complexity

4 Term An entity refers to collection of similar An class refers to group of objects having common
items having same definition. relationships, behaviors and properties.

5 Data RDBMS handles only data. OODBMS handles both data and functions
Handling operating on that data.

6 Objective To keep data independent from To implement data encapsulation.


application program.

7 Key A primary key identifies in object in a Object Id, OID represents an object uniquely in
table uniquely. group of objects.
Data Warehousing
A data warehouse is constructed by integrating data from multiple heterogeneous sources. It
supports analytical reporting, structured and/or ad hoc queries and decision making.
Understanding a Data Warehouse
 A data warehouse is a database, which is kept separate from the organization's
operational database.
 There is no frequent updating done in a data warehouse.
 It possesses consolidated historical data, which helps the organization to analyze its
business.
 A data warehouse helps executives to organize, understand, and use their data to take
strategic decisions.
 Data warehouse systems help in the integration of diversity of application systems.
 A data warehouse system helps in consolidated historical data analysis.

Data Warehouse Features


The key features of a data warehouse are discussed below −
 Subject Oriented − A data warehouse is subject oriented because it provides
information around a subject rather than the organization's ongoing operations. These
subjects can be product, customers, suppliers, sales, revenue, etc. A data warehouse does
not focus on the ongoing operations, rather it focuses on modelling and analysis of data
for decision making.
 Integrated − A data warehouse is constructed by integrating data from heterogeneous
sources such as relational databases, flat files, etc. This integration enhances the
effective analysis of data.
 Time Variant − The data collected in a data warehouse is identified with a particular
time period. The data in a data warehouse provides information from the historical point
of view.
 Non-volatile − Non-volatile means the previous data is not erased when new data is
added to it. A data warehouse is kept separate from the operational database and
therefore frequent changes in operational database is not reflected in the data warehouse.

Data Warehouse Applications


As discussed before, a data warehouse helps business executives to organize, analyze, and use
their data for decision making. A data warehouse serves as a sole part of a plan-execute-assess
"closed-loop" feedback system for the enterprise management. Data warehouses are widely
used in the following fields −

 Financial services
 Banking services
 Consumer goods
 Retail sectors
 Controlled manufacturing

Types of Data Warehouse


Information processing, analytical processing, and data mining are the three types of data
warehouse applications that are discussed below −
 Information Processing − A data warehouse allows to process the data stored in it. The
data can be processed by means of querying, basic statistical analysis, reporting using
crosstabs, tables, charts, or graphs.
 Analytical Processing − A data warehouse supports analytical processing of the
information stored in it. The data can be analyzed by means of basic OLAP operations,
including slice-and-dice, drill down, drill up, and pivoting.
 Data Mining − Data mining supports knowledge discovery by finding hidden patterns
and associations, constructing analytical models, performing classification and
prediction. These mining results can be presented using the visualization tools.

What is Data Warehousing?


Data warehousing is the process of constructing and using a data warehouse. A data warehouse
is constructed by integrating data from multiple heterogeneous sources that support analytical
reporting, structured and/or ad hoc queries, and decision making. Data warehousing involves
data cleaning, data integration, and data consolidations.

Using Data Warehouse Information


There are decision support technologies that help utilize the data available in a data warehouse.
These technologies help executives to use the warehouse quickly and effectively. They can
gather data, analyze it, and take decisions based on the information present in the warehouse.
The information gathered in a warehouse can be used in any of the following domains −
 Tuning Production Strategies − The product strategies can be well tuned by
repositioning the products and managing the product portfolios by comparing the sales
quarterly or yearly.
 Customer Analysis − Customer analysis is done by analyzing the customer's buying
preferences, buying time, budget cycles, etc.
 Operations Analysis − Data warehousing also helps in customer relationship
management, and making environmental corrections. The information also allows us to
analyze business operations.

Functions of Data Warehouse Tools and Utilities


The following are the functions of data warehouse tools and utilities −
 Data Extraction − Involves gathering data from multiple heterogeneous sources.
 Data Cleaning − Involves finding and correcting the errors in data.
 Data Transformation − Involves converting the data from legacy format to warehouse
format.
 Data Loading − Involves sorting, summarizing, consolidating, checking integrity, and
building indices and partitions.
 Refreshing − Involves updating from data sources to warehouse.
What is OLAP?
Online Analytical Processing, a category of software tools which provide analysis of data for
business decisions. OLAP systems allow users to analyze database information from multiple
database systems at one time.
The primary objective is data analysis and not data processing.
Example of OLAP
Any Datawarehouse system is an OLAP system. Uses of OLAP are as follows
 A company might compare their mobile phone sales in September with sales in October,
then compare those results with another location which may be stored in a sperate
database.
 Amazon analyzes purchases by its customers to come up with a personalized homepage
with products which likely interest to their customer.
What is OLTP?
Online transaction processing shortly known as OLTP supports transaction-oriented applications
in a 3-tier architecture. OLTP administers day to day transaction of an organization.
The primary objective is data processing and not data analysis
Example of OLTP system
An example of OLTP system is ATM center. Assume that a couple has a joint account with a
bank. One day both simultaneously reach different ATM centers at precisely the same time and
want to withdraw total amount present in their bank account.
However, the person that completes authentication process first will be able to get money. In this
case, OLTP system makes sure that withdrawn amount will be never more than the amount
present in the bank. The key to note here is that OLTP systems are optimized for transactional
superiority instead data analysis.
KEY DIFFERENCE between OLTP and OLAP:
 Online Analytical Processing (OLAP) is a category of software tools that analyze data
stored in a database whereas Online transaction processing (OLTP) supports transaction-
oriented applications in a 3-tier architecture.
 OLAP creates a single platform for all type of business analysis needs which includes
planning, budgeting, forecasting, and analysis while OLTP is useful to administer day to
day transactions of an organization.
 OLAP is characterized by a large volume of data while OLTP is characterized by large
numbers of short online transactions.
 In OLAP, data warehouse is created uniquely so that it can integrate different data
sources for building a consolidated database whereas OLTP uses traditional DBMS.

Below is the difference between OLAP and OLTP in Data Warehouse:

OLTP vs OLAP

Parameters OLTP OLAP

It is an online transactional system. It manages OLAP is an online analysis and data retrieving
Process
database modification. process.

It is characterized by large numbers of short


Characteristic It is characterized by a large volume of data.
online transactions.

OLAP is an online database query management


Functionality OLTP is an online database modifying system.
system.

Method OLTP uses traditional DBMS. OLAP uses the data warehouse.

Insert, Update, and Delete information from


Query Mostly select operations
the database.

Table Tables in OLTP database are normalized. Tables in OLAP database are not normalized.

OLTP and its transactions are the sources of Different OLTP databases become the source of
Source
data. data for OLAP.
Parameters OLTP OLAP

OLTP database must maintain data integrity OLAP database does not get frequently modified.
Data Integrity
constraint. Hence, data integrity is not an issue.

Response time It’s response time is in millisecond. Response time in seconds to minutes.

The data in the OLTP database is always The data in OLAP process might not be
Data quality
detailed and organized. organized.

It helps to control and run fundamental It helps with planning, problem-solving, and
Usefulness
business tasks. decision support.

Operation Allow read/write operations. Only read and rarely write.

Audience It is a market orientated process. It is a customer orientated process.

Queries in this process are standardized


Query Type Complex queries involving aggregations.
and simple.

Complete backup of the data combined with OLAP only need a backup from time to time.
Back-up
incremental backups. Backup is not important compared to OLTP

DB design is application oriented. Example: DB design is subject oriented. Example:


Design Database design changes with industry like Database design changes with subjects like sales,
Retail, Airline, Banking, etc. marketing, purchasing, etc.

It is used by Data critical users like clerk, Used by Data knowledge users like workers,
User type
DBA & Data Base professionals. managers, and CEO.

Designed for analysis of business measures by


Purpose Designed for real time business operations.
category and attributes.

Performance Transaction throughput is the performance


Query throughput is the performance metric.
metric metric

This kind of Database users allows thousands This kind of Database allows only hundreds of
Number of users
of users. users.

It helps to Increase user’s self-service and Help to Increase productivity of the business
Productivity
productivity analysts.

Challenge Data Warehouses historically have been a An OLAP cube is not an open SQL server data
development project which may prove costly warehouse. Therefore, technical knowledge and
Parameters OLTP OLAP

to build. experience is essential to manage the OLAP


server.

It ensures that response to the query is quicker


Process It provides fast result for daily used data.
consistently.

It lets the user create a view with the help of a


Characteristic It is easy to create and maintain.
spreadsheet.

A data warehouse is created uniquely so that it


OLTP is designed to have fast response time,
Style can integrate different data sources for building a
low data redundancy and is normalized.
consolidated database

Data Mart
Data marts contain a subset of organization-wide data that is valuable to specific groups of
people in an organization. In other words, a data mart contains only those data that is specific to
a particular group. For example, the marketing data mart may contain only data related to items,
customers, and sales. Data marts are confined to subjects.
What is Data Mart?
A Data Mart is focused on a single functional area of an organization and contains a subset of
data stored in a Data Warehouse. A Data Mart is a condensed version of Data Warehouse and is
designed for use by a specific department, unit or set of users in an organization.
E.g., Marketing, Sales, HR or finance. It is often controlled by a single department in an
organization.
Data Mart usually draws data from only a few sources compared to a Data warehouse. Data
marts are small in size and are more flexible compared to a Data warehouse.

The following figures show a graphical representation of data marts.


Why do we need Data Mart?
 Data Mart helps to enhance user’s response time due to reduction in volume of data
 It provides easy access to frequently requested data.
 Data mart are simpler to implement when compared to corporate Data warehouse. At the
same time, the cost of implementing Data Mart is certainly lower compared with
implementing a full data warehouse.
 Compared to Data Warehouse, a data mart is agile. In case of change in model, data mart
can be built quicker due to a smaller size.
 A Data mart is defined by a single Subject Matter Expert. On the contrary data warehouse
is defined by interdisciplinary SME from a variety of domains. Hence, Data mart is more
open to change compared to Data warehouse.
 Data is partitioned and allows very granular access control privileges.
 Data can be segmented and stored on different hardware/software platforms.
Types of Data Mart
There are three main types of data mart:
1. Dependent: Dependent data marts are created by drawing data directly from operational, external or
both sources.
2. Independent: Independent data mart is created without the use of a central data warehouse.
3. Hybrid: This type of data marts can take data from data warehouses or operational systems.
1. Dependent Data Mart
begins with a common source, but they are
scrapped, and mostly junked.
A dependent data mart allows sourcing
organization’s data from a single Data
Warehouse. It is one of the data mart
example which offers the benefit of
centralization. If you need to develop one or
more physical data marts, then you need to
configure them as dependent data marts.
Dependent Data Mart in data warehouse can
be built in two different ways. Either where
a user can access both the data mart and data
warehouse, depending on need, or where
access is limited only to the data mart. The
second approach is not optimal as it
produces sometimes referred to as a data
junkyard. In the data junkyard, all data

Dependent Data Mart


2. Independent Data Mart
An independent data mart is created without
the use of central Data warehouse. This kind
of Data Mart is an ideal option for smaller
groups within an organization.
An independent data mart has neither a
relationship with the enterprise data
warehouse nor with any other data mart. In
Independent data mart, the data is input
separately, and its analyses are also
performed autonomously.
Implementation of independent data marts is
antithetical to the motivation for building a
data warehouse. First of all, you need a
consistent, centralized store of enterprise
data which can be analyzed by multiple
users with different interests who want
widely varying information. Independent Data Mart

3. Hybrid Data Mart:


A hybrid data mart combines input from
sources apart from Data warehouse. This
could be helpful when you want ad-hoc
integration, like after a new group or product
is added to the organization.
It is the best data mart example suited for
multiple database environments and fast
implementation turnaround for any
organization. It also requires least data
cleansing effort. Hybrid Data mart also
supports large storage structures, and it is
best suited for flexible for smaller data-
centric applications.

Hybrid Data Mart


Steps in Implementing a Data mart
Implementing a Data Mart is a rewarding but complex procedure. Here are the detailed steps to
implement a Data Mart:
Designing
Designing is the first phase of Data Mart implementation. It covers all the tasks between
initiating the request for a data mart to gathering information about the requirements. Finally, we
create the logical and physical Data Mart design.
The design step involves the following tasks:
 Gathering the business & technical requirements and Identifying data sources.
 Selecting the appropriate subset of data.
 Designing the logical and physical structure of the data mart.
Data could be partitioned based on following criteria:
 Date
 Business or Functional Unit
 Geography
 Any combination of above
Data could be partitioned at the application or DBMS level. Though it is recommended to
partition at the Application level as it allows different data models each year with the change in
business environment.

What Products and Technologies Do You Need?


A simple pen and paper would suffice. Though tools that help you create UML or ER
diagram would also append meta data into your logical and physical designs.
Constructing
This is the second phase of implementation. It involves creating the physical database and the
logical structures.
This step involves the following tasks:
 Implementing the physical database designed in the earlier phase. For instance, database
schema objects like table, indexes, views, etc. are created.
What Products and Technologies Do You Need?
You need a relational database management system to construct a data mart. RDBMS have
several features that are required for the success of a Data Mart.
 Storage management: An RDBMS stores and manages the data to create, add, and
delete data.
 Fast data access: With a SQL query you can easily access data based on certain
conditions/filters.
 Data protection: The RDBMS system also offers a way to recover from system failures
such as power failures. It also allows restoring data from these backups incase of the disk
fails.
 Multiuser support: The data management system offers concurrent access, the ability
for multiple users to access and modify data without interfering or overwriting changes
made by another user.
 Security: The RDMS system also provides a way to regulate access by users to objects
and certain types of operations.
Populating:
In the third phase, data in populated in the data mart.
The populating step involves the following tasks:
 Source data to target data Mapping
 Extraction of source data
 Cleaning and transformation operations on the data
 Loading data into the data mart
 Creating and storing metadata
What Products and Technologies Do You Need?
You accomplish these population tasks using an ETL (Extract Transform Load) Tool. This tool
allows you to look at the data sources, perform source-to-target mapping, extract the data,
transform, cleanse it, and load it back into the data mart.
In the process, the tool also creates some metadata relating to things like where the data came
from, how recent it is, what type of changes were made to the data, and what level of
summarization was done.
Accessing
Accessing is a fourth step which involves putting the data to use: querying the data, creating
reports, charts, and publishing them. End-user submit queries to the database and display the
results of the queries
The accessing step needs to perform the following tasks:
 Set up a meta layer that translates database structures and objects names into business
terms. This helps non-technical users to access the Data mart easily.
 Set up and maintain database structures.
 Set up API and interfaces if required
What Products and Technologies Do You Need?
You can access the data mart using the command line or GUI. GUI is preferred as it can easily
generate graphs and is user-friendly compared to the command line.
Managing
This is the last step of Data Mart Implementation process. This step covers management tasks
such as-
 Ongoing user access management.
 System optimizations and fine-tuning to achieve the enhanced performance.
 Adding and managing fresh data into the data mart.
 Planning recovery scenarios and ensure system availability in the case when the system
fails.

Best practices for Implementing Data Marts


Following are the best practices that you need to follow while in the Data Mart Implementation
process:
 The source of a Data Mart should be departmentally structured
 The implementation cycle of a Data Mart should be measured in short periods of time,
i.e., in weeks instead of months or years.
 It is important to involve all stakeholders in planning and designing phase as the data
mart implementation could be complex.
 Data Mart Hardware/Software, Networking and Implementation costs should be
accurately budgeted in your plan
 Even though if the Data mart is created on the same hardware they may need some
different software to handle user queries. Additional processing power and disk storage
requirements should be evaluated for fast user response
 A data mart may be on a different location from the data warehouse. That’s why it is
important to ensure that they have enough networking capacity to handle the Data
volumes needed to transfer data to the data mart.
 Implementation cost should budget the time taken for Datamart loading process. Load
time increases with increase in complexity of the transformations.

Advantages of a Data Mart

 Data marts contain a subset of organization-wide data. This Data is valuable to a specific
group of people in an organization.
 It is cost-effective alternatives to a data warehouse, which can take high costs to build.
 Data Mart allows faster access of Data.
 Data Mart is easy to use as it is specifically designed for the needs of its users. Thus a
data mart can accelerate business processes.
 Data Marts needs less implementation time compare to Data Warehouse systems. It is
faster to implement Data Mart as you only need to concentrate the only subset of the data.
 It contains historical data which enables the analyst to determine data trends.
Disadvantages of a Data Mart
 Many a times enterprises create too many disparate and unrelated data marts without
much benefit. It can become a big hurdle to maintain.
 Data Mart cannot provide company-wide data analysis as their data set is limited.

Differences between Data Warehouse and Data Mart

Parameter Data Warehouse Data Mart

Definition A Data Warehouse is a large repository of A data mart is an only subtype of a Data
data collected from different organizations Warehouse. It is designed to meet the need of a
or departments within a corporation. certain user group.

Usage It helps to take a strategic decision. It helps to take tactical decisions for the business.

Objective The main objective of Data Warehouse is A data mart mostly used in a business division at
to provide an integrated environment and the department level.
coherent picture of the business at a point
in time.

Designing The designing process of Data Warehouse The designing process of Data Mart is easy.
is quite difficult.

May or may not use in a dimensional It is built focused on a dimensional model using a
model. However, it can feed dimensional start schema.
models.

Data Handling Data warehousing includes large area of Data marts are easy to use, design and implement
the corporation which is why it takes a as it can only handle small amounts of data.
long time to process it.

Focus Data warehousing is broadly focused all Data Mart is subject-oriented, and it is used at a
the departments. It is possible that it can department level.
even represent the entire company.

Data type The data stored inside the Data Warehouse Data Marts are built for particular user groups.
are always detailed when compared with Therefore, data short and limited.
data mart.

Subject-area The main objective of Data Warehouse is Mostly hold only one subject area- for example,
to provide an integrated environment and Sales figure.
coherent picture of the business at a point
in time.

Data storing Designed to store enterprise-wide decision Dimensional modeling and star schema design
data, not just marketing data. employed for optimizing the performance of
access layer.

Data type Time variance and non-volatile design are Mostly includes consolidation data structures to
strictly enforced. meet subject area's query and reporting needs.

Data value Read-Only from the end-users standpoint. Transaction data regardless of grain fed directly
from the Data Warehouse.

Scope Data warehousing is more helpful as it can Data mart contains data, of a specific department
bring information from any department. of a company. There are maybe separate data
marts for sales, finance, marketing, etc. Has
limited usage

Source In Data Warehouse Data comes from many In Data Mart data comes from very few sources.
sources.

Size The size of the Data Warehouse may range The Size of Data Mart is less than 100 GB.
from 100 GB to 1 TB+.

Implementation The implementation process of Data The implementation process of Data Mart is
time Warehouse can be extended from months restricted to few months.
to years.

Data Mining

Data Mining is defined as the procedure of extracting information from huge sets of data. In
other words, we can say that data mining is mining knowledge from data.

What is Data Mining?


Data mining is looking for hidden, valid, and potentially useful patterns in huge data sets. Data
Mining is all about discovering unsuspected/ previously unknown relationships amongst the data.
It is a multi-disciplinary skill that uses machine learning, statistics, AI and database technology.
The insights derived via Data Mining can be used for marketing, fraud detection, and scientific
discovery, etc.
Data mining is also called as Knowledge discovery, Knowledge extraction, data/pattern analysis,
information harvesting, etc.
Types of Data
Data mining can be performed on following types of data
 Relational databases  Transactional and Spatial databases
 Data warehouses  Heterogeneous and legacy databases
 Advanced DB and information repositories  Multimedia and streaming database
 Object-oriented and object-relational  Text databases
databases  Text mining and Web mining

Data Mining Implementation Process

Let's study the Data Mining implementation process in detail


Business understanding:
In this phase, business and data-mining goals are established.
 First, you need to understand business and client objectives. You need to define what
your client wants (which many times even they do not know themselves)
 Take stock of the current data mining scenario. Factor in resources, assumption,
constraints, and other significant factors into your assessment.
 Using business objectives and current scenario, define your data mining goals.
 A good data mining plan is very detailed and should be developed to accomplish both
business and data mining goals.
Data understanding:
In this phase, sanity check on data is performed to check whether its appropriate for the data
mining goals.
 First, data is collected from multiple data sources available in the organization.
 These data sources may include multiple databases, flat filer or data cubes. There are
issues like object matching and schema integration which can arise during Data
Integration process. It is a quite complex and tricky process as data from various sources
unlikely to match easily. For example, table A contains an entity named cust_no whereas
another table B contains an entity named cust-id.
 Therefore, it is quite difficult to ensure that both of these given objects refer to the same
value or not. Here, Metadata should be used to reduce errors in the data integration
process.
 Next, the step is to search for properties of acquired data. A good way to explore the data
is to answer the data mining questions (decided in business phase) using the query,
reporting, and visualization tools.
 Based on the results of query, the data quality should be ascertained. Missing data if any
should be acquired.
Data preparation:
In this phase, data is made production ready.
The data preparation process consumes about 90% of the time of the project.
The data from different sources should be selected, cleaned, transformed, formatted,
anonymized, and constructed (if required).
Data cleaning is a process to "clean" the data by smoothing noisy data and filling in missing
values.
For example, for a customer demographics profile, age data is missing. The data is incomplete
and should be filled. In some cases, there could be data outliers. For instance, age has a value
300. Data could be inconsistent. For instance, name of the customer is different in different
tables.
Data transformation operations change the data to make it useful in data mining. Following
transformation can be applied

Data Transformation:
Data transformation operations would contribute toward the success of the mining process.
Smoothing: It helps to remove noise from the data.
Aggregation: Summary or aggregation operations are applied to the data. I.e., the weekly sales
data is aggregated to calculate the monthly and yearly total.
Generalization: In this step, Low-level data is replaced by higher-level concepts with the help
of concept hierarchies. For example, the city is replaced by the county.
Normalization: Normalization performed when the attribute data are scaled up o scaled down.
Example: Data should fall in the range -2.0 to 2.0 post-normalization.
Attribute construction: these attributes are constructed and included the given set of attributes
helpful for data mining.
The result of this process is a final data set that can be used in modeling.

Modeling
In this phase, mathematical models are used to determine data patterns.
 Based on the business objectives, suitable modeling techniques should be selected for the
prepared dataset.
 Create a scenario to test check the quality and validity of the model.
 Run the model on the prepared dataset.
 Results should be assessed by all stakeholders to make sure that model can meet data
mining objectives.
Evaluation:
In this phase, patterns identified are evaluated against the business objectives.
 Results generated by the data mining model should be evaluated against the business
objectives.
 Gaining business understanding is an iterative process. In fact, while understanding, new
business requirements may be raised because of data mining.
 A go or no-go decision is taken to move the model in the deployment phase.
Deployment:
In the deployment phase, you ship your data mining discoveries to everyday business operations.
 The knowledge or information discovered during data mining process should be made
easy to understand for non-technical stakeholders.
 A detailed deployment plan, for shipping, maintenance, and monitoring of data mining
discoveries is created.
 A final project report is created with lessons learned and key experiences during the
project. This helps to improve the organization's business policy.
Data Mining Techniques

1. Classification:
This analysis is used to retrieve important
and relevant information about data, and
metadata. This data mining method helps to
classify data in different classes.
2. Clustering:
Clustering analysis is a data mining
technique to identify data that are like each
other. This process helps to understand the
differences and similarities between the
data.
3. Regression:
Regression analysis is the data mining method of identifying and analyzing the relationship
between variables. It is used to identify the likelihood of a specific variable, given the
presence of other variables.
4. Association Rules:
This data mining technique helps to find the association between two or more Items. It
discovers a hidden pattern in the data set.
5. Outer detection:
This type of data mining technique refers to observation of data items in the dataset which do
not match an expected pattern or expected behavior. This technique can be used in a variety
of domains, such as intrusion, detection, fraud or fault detection, etc. Outer detection is also
called Outlier Analysis or Outlier mining.
6. Sequential Patterns:
This data mining technique helps to discover or identify similar patterns or trends in
transaction data for certain period.
7. Prediction:
Prediction has used a combination of the other data mining techniques like trends, sequential
patterns, clustering, classification, etc. It analyzes past events or instances in a right sequence
for predicting a future event.

Challenges of Implementation of Data mine:


 Skilled Experts are needed to formulate the data mining queries.
 Overfitting: Due to small size training database, a model may not fit future states.
 Data mining needs large databases which sometimes are difficult to manage
 Business practices may need to be modified to determine to use the information
uncovered.
 If the data set is not diverse, data mining results may not be accurate.
 Integration information needed from heterogeneous databases and global information
systems could be complex
Data mining Examples:
Example 1:
Consider a marketing head of telecom service provides who wants to increase revenues of
long distance services. For high ROI on his sales and marketing efforts customer profiling is
important. He has a vast data pool of customer information like age, gender, income, credit
history, etc. But its impossible to determine characteristics of people who prefer long distance
calls with manual analysis. Using data mining techniques, he may uncover patterns between
high long distance call users and their characteristics.
For example, he might learn that his best customers are married females between the age of
45 and 54 who make more than $80,000 per year. Marketing efforts can be targeted to such
demographic.
Example 2:
A bank wants to search new ways to increase revenues from its credit card operations. They
want to check whether usage would double if fees were halved.
Bank has multiple years of record on average credit card balances, payment amounts, credit
limit usage, and other key parameters. They create a model to check the impact of the
proposed new business policy. The data results show that cutting fees in half for a targetted
customer base could increase revenues by $10 million.

Benefits of Data Mining:


 Data mining technique helps companies to get knowledge-based information.
 Data mining helps organizations to make the profitable adjustments in operation and
production.
 The data mining is a cost-effective and efficient solution compared to other statistical
data applications.
 Data mining helps with the decision-making process.
 Facilitates automated prediction of trends and behaviors as well as automated
discovery of hidden patterns.
 It can be implemented in new systems as well as existing platforms
 It is the speedy process which makes it easy for the users to analyze huge amount of
data in less time.
Disadvantages of Data Mining
 There are chances of companies may sell useful information of their customers to
other companies for money. For example, American Express has sold credit card
purchases of their customers to the other companies.
 Many data mining analytics software is difficult to operate and requires advance
training to work on.
 Different data mining tools work in different manners due to different algorithms
employed in their design. Therefore, the selection of correct data mining tool is a very
difficult task.
 The data mining techniques are not accurate, and so it can cause serious consequences
in certain conditions.

Data Mining Applications


Applications Usage

Communications Data mining techniques are used in communication sector to predict customer behavior to offer
highly targetted and relevant campaigns.

Insurance Data mining helps insurance companies to price their products profitable and promote new
offers to their new or existing customers.

Education Data mining benefits educators to access student data, predict achievement levels and find
students or groups of students which need extra attention. For example, students who are weak
in maths subject.
Manufacturing With the help of Data Mining Manufacturers can predict wear and tear of production assets.
They can anticipate maintenance which helps them reduce them to minimize downtime.

Banking Data mining helps finance sector to get a view of market risks and manage regulatory
compliance. It helps banks to identify probable defaulters to decide whether to issue credit
cards, loans, etc.
Retail Data Mining techniques help retail malls and grocery stores identify and arrange most sellable
items in the most attentive positions. It helps store owners to comes up with the offer which
encourages customers to increase their spending.
Service Providers Service providers like mobile phone and utility industries use Data Mining to predict the
reasons when a customer leaves their company. They analyze billing details, customer service
interactions, complaints made to the company to assign each customer a probability score and
E-Commerce E-commerce websites use Data Mining to offer cross-sells and up-sells through their websites.
offers incentives.
One of the most famous names is Amazon, who use Data mining techniques to get more
customers into their eCommerce store.
Super Markets Data Mining allows supermarket's develope rules to predict if their shoppers were likely to be
expecting. By evaluating their buying pattern, they could find woman customers who are most
likely pregnant. They can start targeting products like baby powder, baby shop, diapers and so
Crime Data
on. Mining helps crime investigation agencies to deploy police workforce (where is a crime
Investigation most likely to happen and when?), who to search at a border crossing etc.

Bioinformatics Data Mining helps to mine biological data from massive datasets gathered in biology and
medicine.
Unit IV - Integrated Systems, Security and Control
Syllabus: Knowledge based decision support systems, Integrating social media and mobile
technologies in Information system, Security, IS Vulnerability, Disaster Management, Computer
Crimes, Securing the Web.
Knowledge-Based Decision Support System
Implementing knowledge-based decision support system is one of the best ways to capture,
process and store and share knowledge among employees. The information can be easily
accessed by the user to resolve a variety of problems, issues or concerns.

Before the development of knowledge-driven DSS, employees with high intellect had to
perform knowledge-intensive tasks. An expert in a particular area would know how to
approach a problem and go about it. Similarly, knowledge-based DSS asks relevant
questions, offers suggestions and gives advice to solve a problem. The only difference is that
it’s automated and speeds up the whole process.

What is a Knowledge-Based DSS?


A knowledge-based DSS
 is a computer-based reasoning system
 that provides information, comprehension and suggestions to users
 to support them in decision-making.
It’s an integration of computerized business intelligence tools and technologies customized to
the needs and requirements of an organization. So, the focus is on
 Identifying specific knowledge sharing and distribution needs of a company
 Setting objectives that need to be attained with a knowledge-driven DSS
 The selection of appropriate tools and technologies
 Understanding the nature of work and decision-making performed by its potential
users
 Selecting data mining techniques
Key Terms and Concepts
A computer-based reasoning system is similar to any other type of decision support system
when it comes to their architecture. But it turns into a knowledge-drive decision support
system when artificial intelligence technologies, management expert systems, data mining
capabilities and other communication mechanisms are integrated.

Before we dig deeper, let’s learn about few important terms and concepts used alongside
knowledge-drive decision support system. This will help gain an in-depth understanding of
such support systems.

1. Expertise: A knowledge-based DSS comes with a specific problem-solving expertise.


This expertise is based upon three components:
 Knowledge in a particular domain and associated symptoms and signs
 Understanding of the relationships between varied symptoms of a problems
 Skills, ways or methods of solving the problem
2. Expert System: A computer system that imitates the decision making capability of a
human expert is called an expert system or an artificial intelligence system. It is
designed to solve problems by
 Using if-then rules
 Reasoning about knowledge
 Drawing inferences from facts and rules
3. Knowledge Discovery and Data Mining: These are interrelated terms used for the
process of extracting valuable knowledge and discovering patterns, in order to
transform the knowledge into easily comprehendible structure for further use. Data
mining is a buzzword but a misnomer. This is because data mining is a process of
collection, storing and analysis of data and not finding patterns. Knowledge discovery
goes through a series of steps:
 Selection
 Pre-processing
 Transformation
 Data mining
 Interpretation
4. Development Environment: It’s the environment in which a decision support system
is developed. It typically includes software for creating a DSS and knowledge base.
The development environment may vary in size, depending upon
production/development needs.
5. Domain Expert: A domain expert is a subject matter expert who has
expertise/authority in a particular domain. A domain expert is an integral part of the
team working on developing a decision support system.
6. Knowledge Engineer: A technical expert who integrates knowledge into a computer
system when developing a decision support system, in order to solve complex
problems that require human expertise.
7. Knowledge Acquisition: It is extraction/mining of knowledge from various sources,
such as experts, databases and external programs.
8. Knowledge Base: It is the collection and storage of structured (facts, rules,
regulations, characteristics, functions, procedures and relationships) and unstructured
information that will be used by a DSS in decision making.
9. Interface Engine: It is a software system to simplify the conception and development
of application interfaces between application systems. Typically, it’s a middleware
application to transform, route and translate messages between various
communication points.
10. Heuristic: It’s an approach to discovery and problem solving by employing practical
methods. These methods may not be optimal but can help achieve immediate goals.

It’s important to be familiar with technical jargons that experts in this field use, in order to
gain a deeper understanding of knowledge-based DSS.

Characteristics of Knowledge-based Decision Support Systems


A knowledge-based DSS is different from conventional systems in the way knowledge is
extracted, processed and presented. The former attempts to emulate human reasoning while
the latter responses to an even in a predefined manner. The main characteristics of
knowledge-driven decision support systems are:

 These systems aid managers in solving complex problems.


 These systems allow users to interact with them during the process of decision
making.
 The recommendations made by these systems are based on human knowledge.
 These systems use knowledge base that’s engineered keeping in mind the nature of
problems they will solve.
 These systems aid in performing limited tasks.
 These systems use heuristic technique of problem solving.
Developing Knowledge-Based Decision Support Systems
Knowledge-based decision support systems are expert systems that are developed when
decision-making cannot be supported using traditional methods. A knowledge-based DSS
project goes through various stages and can be difficult to manage. It’s important to be
committed to monitor the development of a knowledge-based DSS.

Development Stages

 Domain identification (Choosing a subject matter)


 Conceptualization (idea formation, feasibility testing and commencement)
 Formalization (beginning with development officially)
 Implementation (completion and execution)
 Testing (fixing errors and modifications)
It’s important to monitor project development throughout very closely. It’s a collective effort
of knowledge engineers, domain experts, DSS analysts, users and programmers. And a
project manager keeps track of the scope, time, quality and budget, to ensure optimum
allocation of resources and creation of a quality product. A project manager is a person
responsible accomplishing the pre-decided objectives of a project.

Knowledge-Based DSS Examples


Here are few examples of successful and popular knowledge-based decision support systems:

 TAXADVISOR: As the name suggests, TAXADVISOR assisted attorneys by


collecting client data and suggested actions that clients need to take to settle their
financial profile. Its job was to aid attorneys in taxation and estate planning for clients
with estates greater than $175,000.
 Life Insurance Selection Expert System: The expert system helped Meiji Mutual
Life Insurance Company, one of the oldest insurance companies in Japan, deploys
XpertRule to select the most suitable insurance product for an individual from among
myriads of products.
Integrating Social Medias and Mobile Technology in IS
Social and mobile are two of the largest growth areas in the technology world. Both are
focused on communication, both are undeniably personal and both are very immediate. There
is great synergy between the two areas.
What Is Social Media?
The term social media refers to a computer-based technology that facilitates the sharing of
ideas, thoughts, and information through virtual networks and communities. Social media is
internet-based and gives users quick electronic communication of content, such as personal
information, documents, videos, and photos. Users engage with social media via a computer,
tablet, or smartphone via web-based software or applications. More than 4.5 billion people
use social media, as of October 2021.
Social media originated as a way to interact with friends and family but was later adopted by
businesses that wanted to take advantage of a popular new communication method to reach
out to customers. According to the Pew Research Center, social media users tend to be
younger. Nearly 90% of people between the ages of 18 and 29 used at least one form of
social media.
Types or forms of Social Media
Social media may take the form of a variety of tech-enabled activities. These activities
include photo sharing, blogging, social gaming, social networks, video networks, virtual
worlds, reviews, and much more. Even governments and politicians utilize sharing, business
social media to engage with constituents and voters.
Uses of Social Media – in business
For businesses, social media is an indispensable tool. Companies use the platform to find
and engage with customers, drive sales through advertising and promotion, gauge consumer
trends, and offer customer service or support.
Social media's role in helping businesses is significant. It facilitates communication with
customers, enabling the melding of social interactions on e-commerce sites. Its ability to
collect information helps focus on marketing efforts and market research. It helps in
promoting products and services, as it enables the distribution of targeted, timely, and
exclusive sales and coupons to would-be customers. Further, social media can help in
building customer relationships through loyalty programs linked to social media.
The figures for the most popular social media websites as of January 2021 are as follows:
1. Facebook (2.74 billion users)
2. YouTube (2.29 billion users)
3. WhatsApp (2 billion users)
4. Facebook Messenger (1.3 billion users)
5. Instagram (1.22 billion users)
6. Whatsapp (1.21 billion users)
7. TikTok (689 million users)
8. QQ (617 million users)
9. Douyin (600 million users)
10. Sino Weibo (511 million users)
Business applications of social media
In business, social media is used to market products, promote brands, connect to customers and
foster new business. As a communication platform, social media promotes customer feedback
and makes it easy for customers to share their experiences with a company. Businesses can
respond quickly to positive and negative feedback, address customer problems and maintain or
rebuild customer confidence.

Social media is also used for crowdsourcing. That's the practice of using social networking to
gather knowledge, goods or services. Companies use crowdsourcing to get ideas from
employees, customers and the general public for improving products or developing future
products or services.
Examples of business applications of social media include the following:
 Social media analytics. This is the practice of gathering and analyzing data from blogs
and social media websites to assist in making business decisions. The most common use
of social media analytics is to do customer sentiment analysis.
 Social media marketing (SMM). This application increases a company's brand exposure
and customer reach. The goal is to create compelling content that social media users will
share with their social networks. A key components of SMM is social media optimization
(SMO). SMO is a strategy for drawing new visitors to a website. Social media links and
share buttons are added to content and activities are promoted via status updates, tweets
and blogs.
 Social customer relationship marketing. Social CRM is a powerful business tool. For
example, a Facebook page lets people who like a company's brand to like the business's
page. This, in turn, creates ways to communicate, market and network. Social media sites
give users the option to follow conversations about a product or brand to get real-time
market data and feedback.
 Recruiting. Social recruiting has become a key part of employee recruitment strategies.
It is a fast way to reach a lot of potential candidates, both active job seekers and people
who were not thinking about a job change until they say the recruitment post.
 Enterprise social networking. Businesses also use enterprise social networking to
connect people who share similar interests or activities. Public social media platforms let
organizations stay close to customers and make it easy to conduct market research.
Benefits of social media
Social media provides several benefits, including the following:
User visibility. Social platforms let people easily communicate and exchange ideas or
content.
Business and product marketing. These platforms enable businesses to quickly publicize
their products and services to a broad audience. Businesses can also use social media to
maintain a following and test new markets. In some cases, the content created on social
media is the product.
Audience building. Social media helps entrepreneurs and artists build an audience for their
work. In some cases, social media has eliminated the need for a distributor, because anyone
can upload their content and transact business online.
Mobile Technologies

Mobile technology is a type of technology in which a user utilizes a mobile phone to perform
communications-related tasks, such as communicating with friends, relatives, and others. It is
used to send data from one system to another. Portable two-way communications systems,
computing devices, and accompanying networking equipment make up mobile technology.

Definition: Any gadget with internet capabilities that can be accessed from anywhere is
referred to as mobile technology. Smartphones, tablets, some iPods, and laptops already fall
within this category, but this list will undoubtedly grow in the future years.

Mobile technology is becoming increasingly prevalent. Smartphone users have surpassed 3


billion, and the global mobile workforce is expected to reach 1.87 billion by 2022.
Use of Mobile technology
 The incorporation of mobile technology into business has aided telecollaboration. Now,
people could connect from anywhere using mobile technology, and access the papers
and documents they need to complete collaborative work.
 Employees are no longer confined to their desks; they can work from anywhere in the
world.
 Mobile technology can help your company save time and money. Employees who work
from home save thousands on a regular basis.
 Mobile phones eliminate the need for costly technology like landline carrier services.
 Cloud-based services are less expensive than traditional systems. Technology can also
help your company become more flexible and productive.
 Mobile technology has the potential to boost productivity significantly. Mobile
application integration saves an average of 7.5 hours per week per employee. Workers
can also become more productive with the use of smartphones and mobile gadgets.
 Cloud-based mobile technology applications have been seen to be more useful than any
smartphone, particularly in terms of available storage space.

Integration of Social Media in business


Social media integration gives your audience more ways to engage and interact with your brand.
Social media integration gives your audience more ways to engage and interact with your brand.
You’re creating more opportunities for them to share your content and promote your
products/services for you. The best part: It’s easy to do.
Social Media integration
Definition: Social media integration is the act of using social media accounts as an extension of
your marketing strategy. This is typically accomplished two ways:
i. Directing your social media audience to your website
ii. Allowing your social media accounts to be easily accessed on your website
Think about those social media buttons you see on blog posts and web pages. It allows you to
easily share an interesting piece of content without having to copy and paste the URL. That’s a
perfect example of social media integration in action.
Benefits of Integrating Social Media
Social media integration helps accomplish a few key goals, including increasing your brand
reach and awareness. Social media presents a plethora of opportunities for businesses that can
hugely benefit the brands financially and non-financially. 10 most crucial benefits of
integrating social media.

 Reach a wider audience with extended exposure.


 Amplify your website’s interaction and engagement capabilities.
 Gain insights into consumers’ social behavior.
 Channelize traffic by integrating social media on the website and vice-versa.
 Enhance your website vibrancy through creative visual social content.
 Build brand image & reputation through communications.
 Build social proof by promoting User-generated content from social media.
 Increase your conversions and sales.
 Deliver a superlative consumer experience.

Ways to integrate Social media with your business

Your website and social media should work together seamlessly. This helps promote your brand
while boosting traffic to your social media accounts.

Ways to integrate social media onto your business are.

i. Add social sharing links to your blog posts

These are the social share buttons you see at the bottom of most blog posts. They sometimes also
appear at the top. They help increase awareness of your content, while also giving your readers a
seamless way to share your content. The improved user experience will be a boon to your
website.

ii. Add social posts to your website

One great way to spruce up your website while integrating social media is by including a feed of
social media posts on your pages. These are typically live feeds of your social media posts.
However, you can also use a branded hashtag in order to showcase a feed of posts from your
followers and fans.
iii. Create a social login option

Have you ever gone to a website that allowed you to login using your Google, Facebook, or
Twitter account? Those are great examples of social logins!
It’s so much easier to use a social media account to login than creating an entirely new profile,
picking a password, and confirming it on your email—only to have to log in again when you’re
done. Instead, it’s just a few clicks at most and you’re in.

Not only is this a great way to integrate social media on your website, but it’s also the way most
people prefer to login. In fact, one study found that 73% of users prefer to login using their social
media accounts.

iv. Place Social Share Plugins on Product Pages

Placing social share plugins on product pages of an e-commerce website is a trending practice.
This is because it helps generate social presence and conversation about the product on the
social media channel.

A social share plugin allows the e-commerce store to place social buttons on a product page.
These buttons when used by a website visitor allow him to share the product page on the
chosen social channel. Among the many other displayed social media feeds on website, the
product page will also be shared successfully.
The social share plugins must be placed in close proximity to the image of the product or
services to enable the website visitor to ‘see easily’ and share it quickly.

v. Embed Social Media Widget on Website

Embedding a social media feed widget on a website is one of the most lucrative approaches for
displaying your social media hashtag feed.
A social media widget is capable of displaying all your social media feeds together at one
place. The brand generated posts and fan-created content (user-generated content) are all
aggregated and extracted by this tool into visually creative social media feeds. Social widget
work par excellently for corporate events, product launch events, promotional influencer
meets, weddings, brand activations and more.

vi. Shoppable Social Media Walls on Website

You can collect & curate social media content on your website but the highlight is that you can
make your social media content SHOPPABLE. Using the visual commerce solutions from
Taggbox, you can tag products into you social media posts or UGC posts and add buying
option and embed it as Shoppable social media gallery on website. This will help you increase
your possibilities of conversions, build brand reliability & trust, boost engagement, and deliver
a peerless shopping experience.

vii. Display Social Media Videos

Videos grab user attention and let the message sink into his mind. Social media videos are
simply videos that are created and shared on social channels. Such videos provide an easy
social media integration on website. Thus you can embed the social media feed of your videos
on your website effortlessly.

viii. Integrate Social Media Based Commenting System

Commenting tools encourage conversations and allow for human-to-human interaction thus
creating a strong authenticity and reliability of the information.

These ‘commenting systems’ are designed in a way that requires commenters to sync one of
their social media accounts for commenting. This way only a genuine profile is linked to your
identity as a commenter and it helps reduce the presence of trolls and spams online.

ix. Showcase Social Media Reviews

Showcasing social proof is simply integrating social reviews and recommendations of your
former customers on your website for your potential buyers and website visitors to see. Social
proof has been said to be trusted by 79% of customers and it also helps augment sales and
website conversions. You can also ask your audience to post a review using your hashtag and
then this hashtag can be used to create a display a live hashtag feed across any marketing
touchpoint!
Social media integration strategies for email marketing
You can integrate social media into your emails. Doing so will allow your readers to easily and
quickly find your social accounts and follow you.

a) Add social sharing links to your footer


Your emails are the perfect place to add social sharing links. Like your website they can go to
the top or the bottom of your email.

b) Remind subscribers of your social community


One great tactic for social media integration is sending out email blasts showcasing your social
media accounts. This is a great way to invite your subscribers to connect socially by
incentivizing them with the benefits of doing so.

c) Bolster social media campaigns with email blasts


Have a social media giveaway or contest? Or maybe you have an audience poll you want peoples
thoughts on? Maybe you’re trying to collect some user-generated content for a blog post? Email
blasts are a great way to promote them. This is when you send a single email to your entire list
asking them to complete a call to action.

Integrating Mobile Technology in Business

The main ways to integrate mobile into your marketing mix include:

Develop a Mobile friendly Web Site

Developing a mobile friendly website is at the core of a mobile marketing strategy. Make your
web presence more mobile friendly or better able to accommodate content specific to a
campaign.

The following table outlines the usages of a mobile optimized web site, the benefits, what to
measure and how to measure the outcomes.

Uses Benefits
• Creating mobile landing pagesfor • Improves engagement
campaigns • Reduces scrolling andslow downloads
• Aiding mobile search results, • Creates hyper local marketingopportunities
especially if mobile site is content
• Enables campaign integrationwith other tactics or mediums
specific
• Is quantifiable—can be trackedthrough analytic programs
• Adopting location based point of
sale or “instant purchase” • Provides ability to reach customers‘on-the-go’ and create highly
opportunities effective two-way communication
• Engaging in mobile commerce • Provides single greatest returnon investment
Texting (SMS/MMS)
Text messaging (SMS) and multimedia messaging (MMS) on feature and smartphones have
been well received. They are widely used, not only for personal use, but also in business,
entertainment and education. Texting is twice as popular as browsing or apps. SMS/MMS
marketing represents a more selective and therefore cost-effective opportunity for either
driving traffic or engaging response. The basis of SMS marketing is to make an appealing
offer; it can be a powerful direct response tool with many applications.

Uses Benefits
• Customer text reminders • Is quick and easy to implement
• Links to mobile coupons, contests • Enables timely and relevantexchange of ideas
• Sale notifications • Is convenient
• Last minute alert offers • Has a broad reach
• Transactions—placing orders, or • 90% of devices support SMS
donating to charity, by text • Creates loyalty or a “following”
• Product/service support • Has CRM capability
• Appointment confirmations
• Mobile surveys, polls

Using SMS/MMS technology offered through smart phones to reach a wider audience.

Mobile Applications (apps)


Mobile applications or apps are compact software programs that perform specific tasks for the
mobile user. Some apps are pre-installed on the mobile device, and others can be downloaded
from a server (e.g. Apple Apps, Blackberry App World, Android Market) or from individual
business/organization websites.

Uses Benefits
• Productivity • Can be highly targeted
• Entertainment • Enable easy interaction and communications to engage customers
• Utilities (e.g. weather app) and build deeper relationships
• Social share communications • Encourage brand recognition
• Location Search GPS • Offer unique leading edge tools
• Mobile commerce

Mobile Display Advertising


Mobile display advertising is another way to reach a distinct audience incrementally, whether
you advertise on popular mobile sites related to your industry or through mobile ad networks
(similar to an ad agency). Mobile display ads can provide a direct link to your online sales
channels, enabling greater integration and post-click engagement, especially when used in
conjunction with an optimized mobile site.
Mobile display advertising is a form of advertising via mobile (wireless) phones or other mobile
devices. Key elements of mobile advertising include the advertiser, a publisher or mobile ad
network, mobile system operators and mobile devices.

Uses Benefits
• Reaching mobile audience throughbanner • Provides visual display to highlightspecific campaigns
ads that sit above mobile site content • Delivers personal and locationrelevant messages
• Working in conjunction with a lead • Drives response andbrand relationships
generation landing page
• Banner placement “above the fold”offers the best
• Offering local coupons or specialsin exposure for the advertiser and can quickly engagethe
some form, either in your ads, on your mobile user if the message is targeted to them
site or elsewhere, to drawin customers

Quick Response (QR) Codes and Snap Tags


Quick Response Codes (QR Codes) are two-dimensional bar codes that can be scanned and read
by smartphones. Once mobile users scan or snap a picture of your code, they will be directed to
your web page, shown a video, or receive SMS text, etc.
QR Codes (Quick Response Codes) are barcode images that, when scanned by a mobile device,
take the viewer to a website or landing page.
Snap tags are customized QR codes in ring configurations that, when scanned, open an
interactive communication.

Uses Benefits
QR Code Usages: • Afford more opportunity for sharing and
• Addition to business cards for more contact building community
detail and directlinks to social media • Bridge the gap between online andoffline media –
• Inclusion in print ad materials toexpand on product e.g. provide more details than can fit on print
or services materials
• Addition to direct mail, event materials for more • Enable quick call to action – e.g. while store is
detailed info closed, passersby can still get information or
receivediscount offers
• Placement in storefront windows
• All data collected is reported via an analytics
• Placement on premium itemsSnap Tag Usages:
dashboard andrecorded as attributes in a
• Each code ring position can open up to different consumer database which helps in your CRM
campaigns, response requests or unique reporting efforts
data.
• SnapTags can be used across different media
platforms to delivermarketing campaigns customized
for different consumer segments.
Location Based Service/Geotargeting Mobile Advertising
Geotargeting is a way for your mobile or website to display content-specific information
depending on the location of the user. Through a mobile network that uses geographical
positioning on mobile devices, you can target your marketing by behaviour, knowing where
your prospective target is located, and make offers and calls to action accordingly.
Uses Benefits
• Promoting events • Very effective for attracting
• Responding to user requests forthe nearest service or impulse buyers
business, e.g restaurant, ATM machine • Provides convenience and thusbuilds
• Participating in apps that provide users loyalty
navigation toany street address • Ties in nicely with the global
• Making location-based offers to users of apps that allow positioning software on most smart
them to locate “friends” on a map via mobile and feature phones

• Placing mobile ads that appear only when mobile • Location target ads offer more
users are in aparticular area opportunities to reach clients

• Advertising on game apps • Focused ‘local’ strategies can be built


into your marketing mix
• Targeting your Google ads basedon geographical
location or geographical interest

Mobile Commerce asa Business Activity


Mobile commerce, also known as m-commerce, is the conducting of transactional commerce
using a mobile device. Mobile commerce involves new technologies, services and business
models and is quite different from traditional e-commerce. They open the door to many new
applications and services because they accompany users wherever they go. Users can look for a
nearby restaurant, stay in touch with colleagues, or pay for items at a store.

Uses Benefits
• Promoting products or services on mobile • Real-time apps improvecustomer satisfaction
messaging channels, including email, text • Knowing where your users are can help you optimize
messaging,mobile app push notifications, local dealsand shipping logistics
QR Code scanning, and social networks
• Catches impulse buyers
• Texting SMS to deliver product or
• Allows for instant customerservice interaction
promotional information alerts, track
inquiries and receive order status updates • Builds consumer loyalty throughadded convenience
Security
All sorts of data whether it is government, corporate, or personal need high security; however,
some of the data, which belongs to the government defense system, banks, defense research and
development organization, etc. are highly confidential and even small amount of negligence to
these data may cause great damage to the whole nation. Therefore, such data need security at a
very high level.
The protection of information systems against unauthorized access to or modification of
information, whether in storage, processing or transit, and against the denial of service to
authorized users, including those measures necessary to detect, document, and counter such
threats.
Definition: Information systems security refers to the processes and methodologies involved
with keeping information confidential, available and assuring its integrity.

It also refers to:

 Access controls, which prevent unauthorized personnel from entering or accessing a


system.
 Protecting information no matter where that information is, i.e. in transit (such as in an
email) or in a storage area.
 The detection and remediation of security breaches, as well as documenting those events.

Information systems security does not just deal with computer information, but also protecting
data and information in all of its forms.
Objectives of Information System Security
There are 3 main objectives, commonly known as CIA – Confidentiality, Integrity and
Availability. It is also called as the Information Security Triad (CIA).

1. Confidentiality – means information is not disclosed to unauthorized individuals,


entities and process..
2. Integrity – means maintaining accuracy and completeness of data. This means data
cannot be edited in an unauthorized way..
3. Availability – means information must be available when needed..
Security Threats to Information Systems
A computer system threat is anything that leads to loss or corruption of data or physical damage
to the hardware and/or infrastructure.
What is a Security Threat?
Security Threat is defined as a risk that which can potentially harm computer systems and
organization. The cause could be physical such as someone stealing a computer that contains
vital data. The cause could also be non-physical such as a virus attack. we will define a threat as
a potential attack from a hacker that can allow them to gain unauthorized access to a computer
system.
What are Physical Threats?
A physical threat is a potential cause of an incident that may result in loss or physical
damage to the computer systems.
The following list classifies the physical threats into three (3) main categories;
 Internal: The threats include fire, unstable power supply, humidity in the rooms housing
the hardware, etc.
 External: These threats include Lightning, floods, earthquakes, etc.
 Human: These threats include theft, vandalism of the infrastructure and/or hardware,
disruption, accidental or intentional errors.
To protect computer systems from the above mentioned physical threats, an organization must
have physical security control measures.
The following list shows some of the possible measures that can be taken:
 Internal: Fire threats could be prevented by the use of automatic fire detectors and
extinguishers that do not use water to put out a fire. The unstable power supply can be
prevented by the use of voltage controllers. An air conditioner can be used to control the
humidity in the computer room.
 External: Lightning protection systems can be used to protect computer systems against
such attacks. Lightning protection systems are not 100% perfect, but to a certain extent,
they reduce the chances of Lightning causing damage.
 Humans: Threats such as theft can be prevented by use of locked doors and restricted
access to computer rooms.
What are Non-physical threats?
A non-physical threat is a potential cause of an incident that may result in;
 Loss or corruption of system data
 Disrupt business operations that rely on computer systems
 Loss of sensitive information
 Illegal monitoring of activities on computer systems
 Cyber Security Breaches and Others
The non-physical threats are also known as logical threats. The following list is the common
types of non-physical threats;
 Virus  Unauthorized access to computer
 Trojans systems resources such as data
 Worms  Phishing
 Spyware  Other Computer Security Risks
 Key loggers
 Adware
 Denial of Service Attacks
 Distributed Denial of Service
Attacks
To protect computer systems from the above-mentioned threats, an organization must
have logical security measures in place.
To protect against viruses, Trojans, worms, etc. an organization can use anti-virus software.
In additional to the anti-virus software, an organization can also have control measures on the
usage of external storage devices and visiting the website that is most likely to download
unauthorized programs onto the user’s computer.
Unauthorized access to computer system resources can be prevented by the use of
authentication methods. The authentication methods can be, in the form of user ids and strong
passwords, smart cards or biometric, etc.
Intrusion-detection/prevention systems can be used to protect against denial of service
attacks. There are other measures too that can be put in place to avoid denial of service attacks.
Tools for Security
In order to ensure the confidentiality, integrity, and availability of information, organizations can
choose from a variety of tools.
i) Authentication
Authentication can be accomplished by identifying someone through one or more of three
factors: i) something they know, ii) something they have and iii) something they are.
For example, the most common form of authentication today is the user ID and password.
Identifying someone only by something they have, such as a key or a card, can also be
problematic. When that identifying token is lost or stolen, the identity can be easily stolen.
The final factor, something you are, is much harder to compromise. This factor identifies a user
through the use of a physical characteristic, such as an eye-scan or fingerprint. Identifying
someone through their physical characteristics is called biometrics.
ii) Access Control

Once a user has been authenticated, the next step is to ensure that they can only access the
information resources that are appropriate. This is done through the use of access control. Access
control determines which users are authorized to read, modify, add, and/or delete information.
Several different access control models exist.
iii) Encryption

Many times, an organization needs to transmit information over the Internet or transfer it on
external media such as a CD or flash drive. In these cases, even with proper authentication and
access control, it is possible for an unauthorized person to get access to the data.
Encryption is a process of encoding data upon its transmission or storage so that only authorized
individuals can read it. This encoding is accomplished by a computer program, which encodes
the plain text that needs to be transmitted; then the recipient receives the cipher text and decodes
it (decryption). In order for this to work, the sender and receiver need to agree on the method of
encoding so that both parties can communicate properly. Both parties share the encryption key,
enabling them to encode and decode each other’s messages. This is called symmetric key
encryption. This type of encryption is problematic because the key is available in two different
places.
iv) Password Security
It turns out that this single-factor authentication is extremely easy to compromise. Good
password policies must be put in place in order to ensure that passwords cannot be compromised.
Below are some of the more common policies that organizations should put in place.
 Require complex passwords.
 Change passwords regularly.
 Train employees not to give away passwords.
v) Backups
Another essential tool for information security is a comprehensive backup plan for the entire
organization. Not only should the data on the corporate servers be backed up, but individual
computers used throughout the organization should also be backed up. A good backup plan
should consist of several components.
 A full understanding of the organizational information resources.
 Regular backups of all data. Critical data should be backed up daily, while less critical
data could be backed up weekly.
 Offsite storage of backup data sets. It is essential that part of the backup plan is to store
the data in an offsite location.
 Test of data restoration. This will ensure that the process is working and will give the
organization confidence in the backup plan.
Additional concepts related to backup include the following:
 Universal Power Supply (UPS). A UPS is a device that provides battery backup to
critical components of the system, allowing them to stay online longer and/or allowing
the IT staff to shut them down using proper procedures in order to prevent the data loss
that might occur from a power failure.

vi) Firewalls

Another method that an organization should use to increase security on its network is a firewall.
A firewall can exist as hardware or software (or both).
A hardware firewall is a device that is connected to the network and filters the packets based on a
set of rules.
A software firewall runs on the operating system and intercepts packets as they arrive to a
computer.
A firewall protects all company servers and computers by stopping packets from outside the
organization’s network that do not meet a strict set of criteria. A firewall may also be configured
to restrict the flow of packets leaving the organization. This may be done to eliminate the
possibility of employees watching YouTube videos or using Facebook from a company
computer.
vii) Intrusion Detection Systems

Another device that can be placed on the network for security purposes is an intrusion detection
system, or IDS. An IDS does not add any additional security; instead, it provides the
functionality to identify if the network is being attacked. An IDS can be configured to watch for
specific types of activities and then alert security personnel if that activity occurs.
viii) Sidebar: Virtual Private Networks

A VPN allows a user who is outside of a corporate network to take a detour around the firewall
and access the internal network from the outside. Through a combination of software and
security measures, this lets an organization allow limited access to its networks while at the same
time ensuring overall security.
Physical Security

Physical security is the protection of the actual hardware and networking components that store
and transmit information resources. To implement physical security, an organization must
identify all of the vulnerable resources and take measures to ensure that these resources cannot
be physically tampered with or stolen. These measures include the following.
 Locked doors: It may seem obvious, but all the security in the world is useless if an
intruder can simply walk in and physically remove a computing device. High-value
information assets should be secured in a location with limited access.
 Physical intrusion detection: High-value information assets should be monitored through
the use of security cameras and other means to detect unauthorized access to the physical
locations where they exist.
 Secured equipment: Devices should be locked down to prevent them from being stolen.
One employee’s hard drive could contain all of your customer information, so it is
essential that it be secured.
 Environmental monitoring: An organization’s servers and other high-value equipment
should always be kept in a room that is monitored for temperature, humidity, and airflow.
The risk of a server failure rises when these factors go out of a specified range.
 Employee training: One of the most common ways thieves steal corporate information is
to steal employee laptops while employees are traveling. Employees should be trained to
secure their equipment whenever they are away from the office.

Security Policies

Besides the technical controls listed above, organizations also need to implement security
policies as a form of administrative control. In fact, these policies should really be a starting
point in developing an overall security plan. A good information-security policy lays out the
guidelines for employee use of the information resources of the company and provides the
company recourse in the case that an employee violates a policy.
A security policy should be based on the guiding principles of confidentiality, integrity, and
availability.
A good example of a security policy that many will be familiar with is a web use policy. A web
use policy lays out the responsibilities of company employees as they use company resources to
access the Internet. A security policy should also address any governmental or industry
regulations that apply to the organization.

Cyber Security
Cyber security is a well-designed technique to protect computers, networks, different programs,
personal data, etc., from unauthorized access.
How to Secure Data?
Let us now discuss how to secure data. In order to make your security system strong, you need
to pay attention to the following −
 Security Architecture
 Network Diagram
 Security Assessment Procedure
 Security Policies
 Risk Management Policy
 Backup and Restore Procedures
 Disaster Recovery Plan
 Risk Assessment Procedures
Once you have a complete blueprint of the points mentioned above, you can put better security
system to your data and can also retrieve your data if something goes wrong.

Information System Security


IS security refers to measures put in place to protect information system resources from
unauthorized access or being compromised. Security vulnerabilities are weaknesses in a
computer system, software, or hardware that can be exploited by the attacker to gain
unauthorized access or compromise a system.

Let's now look at some of the threats that information system face

Computer Viruses – these are malicious programs as described in the above section. The
threats posed by viruses can be eliminated or the impact minimized by using Anti-Virus software
and following laid down security best practices of an organization.

Unauthorized Access – the standard convention is to use a combination of a username and a


password. Hackers have learnt how to circumvent these controls if the user does not follow
security best practices. Most organizations have added the use of mobile devices such as phones
to provide an extra layer of security.

Data Loss – if the data center caught fire or was flooded, the hardware with the data can be
damaged, and the data on it will be lost. As a standard security best practice, most organizations
keep backups of the data at remote places. The backups are made periodically and are usually put
in more than one remote area.

Biometric Identification – this is now becoming very common especially with mobile devices
such as smartphones. The phone can record the user fingerprint and use it for authentication
purposes. This makes it harder for attackers to gain unauthorized access to the mobile device.
Such technology can also be used to stop unauthorized people from getting access to your
devices.

Hacking - Hacking refers to the misuse of devices like computers, smartphones, tablets, and
networks to cause damage to or corrupt systems, gather information on users, steal data and
documents, or disrupt data-related activity.

Definition: Hacking is an attempt to exploit a computer system or a private network inside a


computer. Simply put, it is the unauthorised access to or control over computer network security
systems for some illicit purpose.

Types of hackers
In general computer parlance, they are classified as white hats, black hats and grey hats hackers.
White hat hackers hack to check their own security systems to make it more hack-proof.
Black hat hackers hack to take control over the system for personal gains. They can destroy,
steal or even prevent authorized users from accessing the system.
Grey hat hackers are curious people who have just about enough computer language skills to
enable them to hack a system to locate potential loopholes
What is Ethical Hacking?
Ethical Hacking is identifying weakness in computer systems and/or computer networks and
coming with countermeasures that protect the weaknesses. Ethical hackers must abide by the
following rules.
 Get written permission from the owner of the computer system and/or computer
network before hacking.
 Protect the privacy of the organization been hacked.
 Transparently report all the identified weaknesses in the computer system to the
organization.
 Inform hardware and software vendors of the identified weaknesses.
Why Ethical Hacking?
 Information is one of the most valuable assets of an organization. Keeping information
secure can protect an organization’s image and save an organization a lot of money.
 Hacking can lead to loss of business for organizations that deal in finance such as PayPal.
Ethical hacking puts them a step ahead of the cyber criminals who would otherwise lead
to loss of business.
What is Security Testing?
Security testing is a testing technique to determine if an information system protects data and
maintains functionality as intended. It also aims at verifying 6 basic principles as listed below:
 Confidentiality  Authorization
 Integrity  Availability
 Authentication  Non-repudiation

Security Testing is a type of Software Testing that uncovers vulnerabilities, threats, risks in a
software application and prevents malicious attacks from intruders. The purpose of Security
Tests is to identify all possible loopholes and weaknesses of the software system which might
result in a loss of information, revenue, repute at the hands of the employees or outsiders of the
Organization.
The goal of security testing is to identify the threats in the system and measure its potential
vulnerabilities, so the system does not stop functioning or is exploited. It also helps in
detecting all possible security risks in the system and help developers in fixing these
problems through coding.
Types of Security Testing:
There are seven main types of security testing as per Open Source Security Testing methodology
manual. They are explained as follows:
 Vulnerability Scanning: This is done through automated software to scan a system
against known vulnerability signatures.
 Security Scanning: It involves identifying network and system weaknesses, and later
provides solutions for reducing these risks. This scanning can be performed for both
Manual and Automated scanning.
 Penetration testing: This kind of testing simulates an attack from a malicious hacker.
This testing involves analysis of a particular system to check for potential vulnerabilities
to an external hacking attempt.
 Risk Assessment: This testing involves analysis of security risks observed in the
organization. Risks are classified as Low, Medium and High. This testing recommends
controls and measures to reduce the risk.
 Security Auditing: This is an internal inspection of Applications and Operating systems
for security flaws. An audit can also be done via line by line inspection of code
 Ethical hacking: It's hacking an Organization Software systems. Unlike malicious
hackers, who steal for their own gains, the intent is to expose security flaws in the
system.
 Posture Assessment: This combines Security scanning, Ethical Hacking and Risk
Assessments to show an overall security posture of an organization.
Vulnerability
Vulnerability is any mistakes or weakness in the system security procedures, design,
implementation or any internal control that may result in the violation of the system's security
policy.
A flaw or weakness in a system's design, implementation, or operation and management that
could be exploited to violate the system's security policy
Vulnerability—Weakness in an information system, system security procedures, internal
controls, or implementation that could be exploited by a threat source
Vulnerabilities are related to:
 physical environment of the system  hardware
 the personnel  software
 management  communication equipment and
 administration procedures and facilities
security measures within the  peripheral devices
organization  their combinations.
 business operation and service
delivery
It is evident that a pure technical approach cannot even protect physical assets: one should have
administrative procedure to let maintenance personnel to enter the facilities and people with
adequate knowledge of the procedures, motivated to follow it with proper care.
Four examples of vulnerability exploits:
 an attacker finds and uses an overflow weakness to install malware to export sensitive
data;
 an attacker convinces a user to open an email message with attached malware;
 an insider copies a hardened, encrypted program onto a thumb drive and cracks it at
home;
 a flood damages one's computer systems installed at ground floor.
Classification of Vulnerabilities
Vulnerabilities are classified according to the asset class they are related to:

 hardware o insecure network architecture


o susceptibility to humidity  personnel
o susceptibility to dust o inadequate recruiting process
o susceptibility to soiling o inadequate security awareness
o susceptibility to unprotected storage  physical site
 software o area subject to flood
o insufficient testing o unreliable power source
o lack of audit trail  organizational
o design flaw o lack of regular audits
 network o lack of continuity plans
o unprotected communication lines o lack of security
Causes of Vulnerability
 Complexity: Large, complex systems increase the probability of flaws and
unintended access points.
 Familiarity: Using common, well-known code, software, operating systems, and/or
hardware increases the probability an attacker has or can find the knowledge and tools to
exploit the flaw
 Connectivity: More physical connections, privileges, ports, protocols, and services and
time each of those are accessible increase vulnerability.
 Password management flaws: The computer user uses weak passwords that could be
discovered by brute force.[19] The computer user stores the password on the computer
where a program can access it. Users re-use passwords between many programs and
websites.
 Fundamental operating system design flaws: The operating system designer chooses to
enforce suboptimal policies on user/program management. For example, operating
systems with policies such as default permit grant every program and every user full
access to the entire computer. This operating system flaw allows viruses and malware to
execute commands on behalf of the administrator.
 Internet Website Browsing: Some internet websites may contain
harmful Spyware or Adware that can be installed automatically on the computer systems.
After visiting those websites, the computer systems become infected and personal
information will be collected and passed on to third party individuals.
 Software bugs: The programmer leaves an exploitable bug in a software program. The
software bug may allow an attacker to misuse an application.
 Unchecked user input: The program assumes that all user input is safe. Programs that do
not check user input can allow unintended direct execution of commands or SQL
statements (known as Buffer overflows, SQL injection or other non-validated inputs).
 Not learning from past mistakes: for example most vulnerabilities discovered
in IPv4 protocol software were discovered in the new IPv6 implementations.
The research has shown that the most vulnerable point in most information systems is the human
user, operator, designer, or other human: so humans should be considered in their different roles
as asset, threat, information resources.
What is Vulnerability Assessment / Testing?
Vulnerability Assessment is a process to evaluate the security risks in the software system in
order to reduce the probability of a threat. It is also called Vulnerability Testing.
The purpose of Vulnerability Assessment is to reduce the possibility for intruders (hackers) to
get unauthorized access.
How to do Vulnerability Testing
Following is the step by step Vulnerability
Assessment Methodology/ Technique
Step 1) Setup:
 Begin Documentation  Update Tools
 Secure Permission  Configure Tools
Step 2) Test Execution:
 Run the Tools
 Run the captured data packet
Step 3) Vulnerability Analysis:
 Defining and classifying network or System resources.
 Assigning priority to the resource( Ex: - High, Medium, Low)
 Identifying potential threats to each resource.
 Developing a strategy to deal with the most prioritize problems first.
 Defining and implementing ways to minimize the consequences if an attack occurs.
Step 4) Reporting
Step 5) Remediation:
 The process of fixing the vulnerabilities.
 For every vulnerability

Advantages of Vulnerability Assessment


 Open Source tools are available.  Automated for Scanning.
 Identifies almost all vulnerabilities  Easy to run on a regular basis.

Disadvantages of Vulnerability Assessment


 High false positive rate
 Can easily detect by Intrusion Detection System Firewall.
 Often fail to notice the latest vulnerabilities.
Vulnerability Testing Methods
Active Testing
 In active Testing, a tester introduces new test data and analyzes the results.
 During the testing process, the testers create a mental model of the process, and it will
grow further during the interaction with the software under test.
 While doing the test, the tester will actively involve in the process of finding out the new
test cases and new ideas. That's why it is called Active Testing.
Passive Testing
 Passive testing, monitoring the result of running software under test without introducing
new test cases or data
Network Testing
 Network Testing is the process of measuring and recording the current state of network
operation over a period of time.
 Testing is mainly done for predicting the network operating under load or to find out the
problems created by new services.
 We need to Test the following Network Characteristics:-
 Utilization levels
 Number of Users
 Application Utilization
Distributed Testing
 Distributed Tests are applied for testing distributed applications, which means, the
applications that are working with multiple clients simultaneously. Basically, testing a
distributed application means testing its client and server parts separately, but by using a
distributed testing method, we can test them all together.
 The test parts will interact with each other during the Test Run. This makes them
synchronized in an appropriate manner. Synchronization is one of the most crucial points
in distributed testing.
Disaster Management
Disaster Management is a strategic planning and procedure that is administered and employed to
protect critical infrastructures (also known as "critical assets") from severe damages when
natural or human made calamities and catastrophic even occur.
Disaster management plans are multi-layered and are aimed to address such issues as floods,
hurricanes, fires, bombings, and even mass failures of utilities or the rapid spread of disease
Though vulnerability to disaster varies, there are four main types of disaster
 Natural disasters. These disasters include floods, hurricanes, earthquakes and volcano
eruptions that can have immediate impacts on human health, as well as secondary
impacts causing further death and suffering from floods causing landslides, earthquakes
resulting in fires, tsunamis causing widespread flooding and typhoons sinking ferries
 Environmental emergencies. These emergencies include technological or industrial
accidents, usually involving hazardous material, and occur where these materials are
produced, used or transported. Large forest fires are generally included in this definition
because they tend to be caused by humans.
 Complex emergencies. These emergencies involve a break-down of authority, looting
and attacks on strategic installations. Complex emergencies include conflict situations
and war.
 Pandemic emergencies. These emergencies involve a sudden onset of a contagious
disease that affects health but also disrupts services and businesses, bringing economic
and social costs.
Emergency Management
Emergency Management is the generic name of an interdisciplinary field dealing with the
strategic organization management processes used to protect asses of an organization from
hazard risks that can cause disasters or catastrophes, and to ensure the continuance of the
organization within their planned lifetime.
Emergency Management is a systematic process leading to action before, during and after a
disaster to save lives and prevent injury.
"Disaster" here means a major emergency that exceeds the community's capacity to respond
successfully with its own resources.
Emergency Management is organized into four phases:
 Mitigation: actions taken to eliminate a hazard or reduce its potential impact.
 Preparedness: planning for major emergencies, including training and exercises.
 Response: actions taken in response to emergencies.
 Recovery: actions taken after a disaster to restore services and reconstruct communities.
Resilience and Prevention
Resilience defines the capability to recover from catastrophe or misfortune. Prevention is totally
avoiding danger or risky events.
Mitigation
Mitigation is the effort to reduce loss of life and property by lessening the impact of disasters.
Preparedness
Preparedness is way of mitigating unwanted outcome and it is one of the crucial actions in
achieving safety and security in the event of calamities, disasters, and terrorism.
Response
An emergency response plan must provide the resources and information needed to evaluate the
human and environmental health impacts of the event, assess and reduce human exposures to
contaminants, and develop science-based strategies for remediation and rebuilding.
Recovery
Recovery efforts are primarily concerned with actions that involve rebuilding destroyed property,
re-employment, and the repair of other essential infrastructure.
Disaster Recovery
Disaster recovery is generally a planning process and it produces a document which ensures
businesses to solve critical events that affect their activities. Such events can be a natural
disaster (earthquakes, flood, etc.), cyber–attack or hardware failure like servers or routers.
As such having a document in place it will reduce the down time of business process from the
technology and infrastructure side. This document is generally combined with Business
Continuity Plan which makes the analyses of all the processes and prioritizes them according to
the importance of the businesses. In case of a massive disruption it shows which process should
be recovered firstly and what should be the downtime. It also minimizes the application service
interruption. It helps us to recover data in the organized process and help the staff to have a
clear view about what should be done in case of a disaster.
Requirements to Have a Disaster Recovery Plan
Disaster recovery starts with an inventory of all assets like computers, network equipment,
server, etc. and it is recommended to register by serial numbers too. We should make an
inventory of all the software and prioritize them according to business importance.
An example is shown in the following table −

Systems Down Time Disaster type Preventions Solution strategy Recover fully

Payroll Server We take Restore the backups in Fix the primary server and
8 hours
system damaged backup daily the Backup Server restore up to date data

You should prepare a list of all contacts of your partners and service providers, like ISP contact
and data, license that you have purchased and where they are purchased. Documenting all your
Network which should include IP schemas, usernames and password of servers.
Preventive steps to be taken for Disaster Recovery
 The server room should have an authorized level. For example: only IT personnel should
enter at any given point of time.
 In the server room there should be a fire alarm, humidity sensor, flood sensor and a
temperature sensor.
 At the server level, RAID systems should always be used and there should always be a
spare Hard Disk in the server room.
 You should have backups in place, this is generally recommended for local and off-site
backup, so a NAS should be in your server room.
 Backup should be done periodically.
 The connectivity to internet is another issue and it is recommended that the headquarters
should have one or more internet lines. One primary and one secondary with a device
that offers redundancy.
 If you are an enterprise, you should have a disaster recovery site which generally is
located out of the city of the main site. The main purpose is to be as a stand-by as in any
case of a disaster, it replicates and backs up the data.
What is Recovery Testing?
Recovery testing verifies the system's ability to recover from points of failure like
software/hardware crashes, network failures etc. The purpose of Recovery Testing is to
determine whether operations can be continued after a disaster or after the integrity of the system
has been lost. It involves reverting to a point where the integrity of the system was known and
then reprocessing transactions up to the point of failure.
Recovery Testing Example
When an application is receiving data from the network, unplug the connecting cable.
 After some time, plug the cable back in and analyze the application’s ability to continue
receiving data from the point at which the network connection was broken.
 Restart the system while a browser has a definite number of sessions open and check
whether the browser is able to recover all of them or not
In Software Engineering, Recoverability Testing is a type of Non- Functional Testing. (Non-
functional testing refers to aspects of the software that may not be related to a specific function
or user action such as scalability or security.)
The time taken to recover depends upon:
 The number of restart points
 A volume of the applications
 Training and skills of people conducting recovery activities and tools available for
recovery.
When there are a number of failures then instead of taking care of all failures, the recovery
testing should be done in a structured fashion which means recovery testing should be carried out
for one segment and then another.
It is done by professional testers. Before recovery testing, adequate backup data is kept in secure
locations. This is done to ensure that the operation can be continued even after a disaster.
Life Cycle of Recovery Process
The life cycle of the recovery process can be classified into the following five steps:

1. Normal operation
2. Disaster occurrence
3. Disruption and failure of the
operation
4. Disaster clearance through the
recovery process
5. Reconstruction of all processes and
information to bring the whole
system to move to normal operation
Let’s discuss these 5 steps in detail-
1. A system consisting of hardware, software, and firmware integrated to achieve a
common goal is made operational for carrying out a well-defined and stated goal. The
system is called to perform the normal operation to carry out the designed job without
any disruption within a stipulated period of time.
2. A disruption may occur due to malfunction of the software, due to various reasons
like input initiated malfunction, software crashing due to hardware failure, damaged
due to fire, theft, and strike.
3. Disruption phase is a most painful phase which leads to business losses, relation
break, opportunity losses, man-hour losses and invariably financial and goodwill
losses. Every sensible agency should have a plan for disaster recovery to enable the
disruption phase to be minimal.
4. If a backup plan and risk mitigation processes are at the right place before
encountering disaster and disruption, then recovery can be done without much loss of
time, effort and energy. A designated individual, along with his team with the
assigned role of each of these persons should be defined to fix the responsibility and
help the organization to save from long disruption period.
5. Reconstruction may involve multiple sessions of operation to rebuild all folders along
with configuration files. There should be proper documentation and process of
reconstruction for correct recovery.
Restoration Strategy
The recovery team should have their unique strategy for retrieving the important code and
data to bring the operation of the agency back to normalcy.
The strategy can be unique to each organization based on the criticality of the systems they
are handling.
The possible strategy for critical systems can be visualized as follows:
1. To have a single backup or more than one
2. To have multiple back-ups at one place or different places
3. To have an online backup or offline backup
4. Can the backup is done automatically based on a policy or to have it manually?
5. To have an independent restoration team or development team itself can be utilized
for the work
Each of these strategies has cost factor associated with it and multiple resources required for
multiple back-ups may consume more physical resources or may need an independent team.
Many companies may be affected due to their data and code dependency on the concerned
developer agency. For instance, if Amazon AWS goes down its shuts 25 of the internet.
Independent Restoration is crucial in such cases.
How to do Recovery Testing
While performing recovery testing following things should be considered.
 We must create a test bed as close to actual conditions of deployment as possible.
Changes in interfacing, protocol, firmware, hardware, and software should be as close
to the actual condition as possible if not the same condition.
 Through exhaustive testing may be time-consuming and a costly affair, identical
configuration, and complete check should be performed.
 If possible, testing should be performed on the hardware we are finally going to
restore. This is especially true if we are restoring to a different machine than the one
that created the backup.
 Some backup systems expect the hard drive to be exactly the same size as the one the
backup was taken from.
 Obsolescence should be managed as drive technology is advancing at a fast pace, and
old drive may not be compatible with the new one. One way to handle the problem is
to restore to a virtual machine. Virtualization software vendors like VMware Inc. can
configure virtual machines to mimic existing hardware, including disk sizes and other
configurations.
 Online backup systems are not an exception for testing. Most online backup service
providers protect us from being directly exposed to media problems by the way they
use fault-tolerant storage systems.
 While online backup systems are extremely reliable, we must test the restore side of
the system to make sure there are no problems with the retrieval functionality,
security or encryption.
Testing procedure after restoration
Most large corporations have independent auditors to perform recovery test exercises
periodically.
The expense of maintaining and testing a comprehensive disaster recovery plan can be
substantial, and it may be prohibitive for smaller businesses.
Smaller risks may rely on their data backups and off-site storage plans to save them in the
case of a catastrophe.
After folders and files are restored, following checks can be done to assure that files are
recovered properly:
 Rename the corrupted document folder
 Count the files in the restored folders and match with it with an existing folder.
 Open a few of the files and make sure they are accessible. Be sure to open them with
the application that normally uses them. And make sure you can browse the data,
update the data or whatever you normally do.
 It is best to open several files of different types, pictures, mp3s, documents and some
large and some small.
 Most operating systems have utilities that you can use to compare files and
directories.

Computer Crimes

Computer crime / Cybercrime is defined as an unlawful action against any person using a
computer, its systems, and its online or offline applications. It occurs when information
technology is used to commit or cover an offense. However, the act is only considered
Cybercrime if it is intentional and not accidental.
The crime that involves and uses computer devices and Internet, is known as Computer or
Cybercrime.
What is Cybercrime?
Cyber-crime is the use of computers and networks to perform illegal activities such as
spreading computer viruses, online bullying, performing unauthorized electronic fund
transfers, etc. Most cybercrimes are committed through the internet. Some cybercrimes can
also be carried out using Mobile phones via SMS and online chatting applications.
Examples of Cybercrime
The following list presents the common types of cybercrimes:
 Computer Fraud: Intentional deception for personal gain via the use of computer
systems.
 Privacy violation: Exposing personal information such as email addresses, phone
number, account details, etc. on social media, websites, etc.
 Identity Theft: Stealing personal information from somebody and impersonating that
person.
 Sharing copyrighted files/information: This involves distributing copyright protected
files such as eBooks and computer programs etc.
 Electronic funds transfer: This involves gaining an un-authorized access to bank
computer networks and making illegal fund transfers.
 Electronic money laundering: This involves the use of the computer to launder money.
 ATM Fraud: This involves intercepting ATM card details such as account number and
PIN numbers. These details are then used to withdraw funds from the intercepted
accounts.
 Denial of Service Attacks: This involves the use of computers in multiple locations to
attack servers with a view of shutting them down.
 Spam: Sending unauthorized emails. These emails usually contain advertisements.
Types of Cybercrime
Let us now discuss the major types of cybercrime −
1) Hacking
It is an illegal practice by which a hacker breaches the computer’s security system of
someone for personal interest.
2) Unwarranted mass-surveillance
Mass surveillance means surveillance of a substantial fraction of a group of people by the
authority especially for the security purpose, but if someone does it for personal interest, it is
considered as cybercrime.
3) Child pornography
It is one of the most heinous crimes that is brazenly practiced across the world. Children are
sexually abused and videos are being made and uploaded on the Internet.
4) Child grooming
It is the practice of establishing an emotional connection with a child especially for the
purpose of child-trafficking and child prostitution.
5) Copyright infringement
If someone infringes someone’s protected copyright without permission and publishes that
with his own name, is known as copyright infringement.
6) Identity theft
Identity theft occurs when a cyber-criminal impersonates someone else identity to practice
malfunction.
7) Denial of Service Attack:
In this cyberattack, the cyber-criminal uses the bandwidth of the victim's network or fills their
e-mail box with spammy mail. Here, the intention is to disrupt their regular services.
8) Software Piracy:
Theft of software by illegally copying genuine programs or counterfeiting. It also includes the
distribution of products intended to pass for the original.
9) Phishing:
Phishing is a technique of extracting confidential information from the bank/financial
institutional account holders by illegal ways.
10) Spoofing:
It is an act of getting one computer system or a network to pretend to have the identity of
another computer. It is mostly used to get access to exclusive privileges enjoyed by that
network or computer.
11) Click fraud
Advertising companies such as Google AdSense offer pay per click advertising services.
Click fraud occurs when a person clicks such a link with no intention of knowing more about
the click but to make more money. This can also be accomplished by using automated
software that makes the clicks.
12) Advance Fee Fraud
An email is sent to the target victim that promises them a lot of money in favor of helping
them to claim their inheritance money.
13) Cyber-extortion
When a hacker hacks someone’s email server, or computer system and demands money to
reinstate the system, it is known as cyber-extortion.
14) Cyber-terrorism
Normally, when someone hacks government’s security system or intimidates government or
such a big organization to advance his political or social objectives by invading the security
system through computer networks, it is known as cyber-terrorism.
15) Computer virus
Viruses are unauthorized programs that can annoy users, steal sensitive data or be used to
control equipment that is controlled by computers.
Securing the Web
There are many ways to assure yourself, employees, and customers that your website is safe.
Website security does not have to be a guessing game. Take essential steps towards
improving your site’s security. Help keep data away from prying eyes.
No method can guarantee your site will forever be “hacker-free.” The use of preventative
methods will reduce your site’s vulnerability. Website security is both a simple and
complicated process.
Website Vulnerabilities & Threats
The most common website security vulnerabilities and threats are:
1. SQL Injections
SQL injection attacks are done by injecting malicious code in a vulnerable SQL query. They
rely on an attacker adding a specially crafted request within the message sent by the website
to the database. A successful attack will alter the database query in such a way that it will
return the information desired by the attacker, instead of the information the website
expected.
2. Cross-site Scripting (XSS)
Cross-site scripting attacks consist of injecting malicious client-side scripts into a website and
using the website as a propagation method.
The danger behind XSS is that it allows an attacker to inject content into a website and
modify how it is displayed, forcing a victim’s browser to execute the code provided by the
attacker when loading the page. If a logged in site administrator loads the code, the script will
be executed with their level of privilege, which could potentially lead to site takeover.
3. Credential Brute Force Attacks
Gaining access to a website’s admin area, control panel or even to the SFTP server is one of
the most common vectors used to compromise websites. The process is very simple; the
attackers basically program a script to try multiple combinations of usernames and passwords
until it finds one that works.
4. Website Malware Infections & Attacks
Using some of the previous security issues as a means to gain unauthorized access to a
website, attackers can then:
 Inject SEO spam on the page
 Drop a backdoor to maintain access
 Collect visitor information or credit card data
 Run exploits on the server to escalate access level
 Use visitors’ computers to mine crypto currencies
 Store botnets command & control scripts
 Show unwanted ads, redirect visitors to scam sites
 Host malicious downloads
 Launch attacks against other sites

5. DoS/DDoS Attacks
A Distributed Denial of Service (DDoS) attack is a non-intrusive internet attack. It is made to
take down the targeted website or slow it down by flooding the network, server or application
with fake traffic.
DDoS attacks are threats that website owners must familiarize themselves with as they are a
critical piece of the security landscape. When a DDoS attack targets a vulnerable resource-
intensive endpoint, even a tiny amount of traffic is enough for the attack to be successful.
Website Security Framework
The US National Institute of Standards and Technology (NIST) developed The Cyber
security Framework which forms the basis of our website security principles framework.
Knowing security is a continuous process means it starting with the foundation of a website
security framework. This framework will involve creating a “culture of security” where
scheduled audits will help in keeping things simple and timely.
The five functions: Identify, Protect, Detect, Respond and Recover will be broken out in
more detail along with actions to be applied.

There are at least ten essential steps you can take to improve website safety.
How to Improve Your Websites Safety
1. Keep Software And Plugins Up-To-Date
Every day, there are countless websites compromised due to outdated software. Potential
hackers and bots are scanning sites to attack. Updates are vital to the health and security of
your website. If your site’s software or applications are not up-to-date, your site is not secure.
Take all software and plugin update requests seriously.
Updates often contain security enhancements and vulnerability repairs. Check your website
for updates or add an update notification plugin. Some platforms allow automatic updates,
which is another option to ensure website security.
2. Add HTTPS and an SSL Certificate
To keep your website safe, you need a secure URL. If your site visitors offer to send their
private information, you need HTTPS, not HTTP, to deliver it.
What is HTTPs?
HTTPS (Hypertext Transfer Protocol Secure) is a protocol used to provide security over the
Internet. HTTPS prevents interceptions and interruptions from occurring while the content is
in transit.
For you to create a secure online connection, your website also needs an SSL Certificate. If
your website asks visitors to register, sign-up, or make a transaction of any kind, you need to
encrypt your connection.
What is SSL?
SSL (Secure Sockets Layer) is another necessary site protocol. This transfers visitor’s
personal information between the website and your database. SSL encrypts information to
prevent it from others reading it while in transit.
It denies those without proper authority the ability to access the data, as well.
3. Choose a Smart Password
With there being so many websites, databases, and programs needing passwords, it is hard to
keep track. A lot of people end up using the same password in all places, to remember their
login information. But this is a significant security mistake.
Create a unique password for every new log in request. Come up with complicated, random,
and difficult to guess passwords. Then, store them outside the website directory.
Refrain from using any personal information inside your password as well. Do not use your
birthday or pet’s name; make it completely unguessable. After three months or sooner,
change password to another one, then repeat. Smart passwords are long and should be at least
twelve characters. Password needs to be a combination of numbers and symbols. Make sure
to alternate between uppercase and lowercase letters. Never use the same password twice or
share it with others.
If you are a business owner or CMS manager, ensure all employees change their passwords
frequently.
4. Use a Secure Web Host
Think of your website’s domain name as a street address. Now, think of the web host as the
plot of “real estate” where your website exists online.
As you would research a plot of land to build a house, you need to examine potential web
hosts to find the right one for you. Many hosts provide server security features that better
protect your uploaded website data. There are certain items to check for when choosing a
host.
 Does the web host offer a Secure File Transfer Protocol (SFTP)?
 Is FTP Use by Unknown User disabled?
 Does it use a Rootkit Scanner?
 Does it offer file backup services?
 How well do they keep up to date on security upgrades?
Whether you choose SiteGround or WP Engine as your web host, make sure it has what you
need to keep your site secure.
5. Record User Access and Administrative Privileges
Initially, you may feel comfortable giving several high-level employees access to your
website. If they make a mistake or overlook an issue, this can result in a significant security
issue. It is vital to vet your employees before giving them website access. Find out if they
have experience using your CMS and if they know what to look for to avoid a security
breach. Educate every CMS user about the importance of passwords and software updates.
Tell them all the ways they can help maintain the website’s safety.
To keep track of who has access to your CMS and their administrative settings, make a
record and update it often. Employees come and go. One of the best ways to prevent security
issues is to have a physical record of who does what with your website.
6. Change Your CMS Default Settings
The most common attacks against websites are entirely automated. What many attack bots
rely on is for users to have their CMS settings on default. After choosing your CMS, change
your default settings immediately. Changes help prevent a large number of attacks from
occurring.
CMS settings can include adjusting control comments, user visibility, and permissions.
Customize users and their permission settings. Do not keep the default settings as is, or you
will run into website security issues at some point.
7. Backup Your Website
One of the best methods to keep your site safe is to have a good backup solution. You should
have more than one. Each is crucial to recovering your website after a major security incident
occurs.
There are several different solutions you can use to help recover damaged or lost files.
Keep your website information off-site. Do not store your backups on the same server as your
website; they are as vulnerable to attacks too.
Choose to keep your website backup on a home computer or hard drive. Find an off-site place
to store your data and to protect it from hardware failures, hacks, and viruses.
Another option is to back up your website in the cloud. It makes storing data easy and allows
access to information from anywhere. Be redundant in your backup process — backup your
backup. By doing this, you can recover files from any point before the hack or virus occurs.
8. Know Your Web Server Configuration Files
Get to know your web server configuration files. You can find them in the root web directory.
Web server configuration files permit you to administer server rules. This includes directives
to improve your website security.
9. Apply for a Web Application Firewall
Make sure you apply for a web application firewall (WAF). It sets between your website
server and the data connection. The purpose is to read every bit of data that passes through it
to protect your site.
10. Tighten Network Security
When you think your website is secure, you need to analyze your network security.
Employees who use office computers may inadvertently be creating an unsafe pathway to
your website.
To prevent them from giving access to your website’s server, consider doing the following at
your business:
o Have computer logins expire after a short period of inactivity.
o Make sure your system notifies users every three months of password changes.
o Ensure all devices plugged into the network are scanned for malware each
time they are attached.
Unit V - New I.T Initiatives
Syllabus: Introduction to Deep learning, Big data, Pervasive Computing, Cloud computing,
Advancements in AI, IoT, Block chain, Crypto currency, Quantum computing

Introduction to Deep Learning


What is Deep Learning?
Deep learning is a branch of machine learning which is completely based on artificial neural
networks, as neural network is going to mimic the human brain. So deep learning is also a
kind of mimic of human brain. In deep learning, we don’t need to explicitly program
everything. Earlier we did not have that much processing power and a lot of data. As in the
last 20 years, the processing power increases exponentially, so deep learning and machine
learning came in the picture.

Deep Learning is a subset of Machine Learning that uses mathematical functions to map the
input to the output. These functions can extract non-redundant information or patterns from the
data, which enables them to form a relationship between the input and the output.
Deep learning is extremely powerful when the dataset is large. It can learn any complex patterns
from the data and can draw accurate conclusions on its own. In fact, deep learning is so powerful
that it can even process unstructured data - data that is not adequately arranged like text corpus,
social media activity, etc. Furthermore, it can also generate new data samples and find anomalies
that machine learning algorithms and human eyes can miss.
In traditional computer programming, input and a set of rules are combined together to get the
desired output. In machine learning and deep learning, input and output are correlated to the
rules.
These rules—when combined with new input—yield desired results.

Deep learning can also be thought of as


an approach to Artificial Intelligence, a
smart combination of hardware and
software to solve tasks requiring human
intelligence.
A Quick History of Deep Learning

Difference between Machine Learning and Deep Learning:


Machine Learning Deep Learning

Works on small amount of Dataset for accuracy. Works on Large amount of Dataset.

Dependent on Low-end Machine. Heavily dependent on High-end Machine.

Divides the tasks into sub-tasks, solves them


individually and finally combine the results. Solves problem end to end.

Takes less time to train. Takes longer time to train.

Testing time may increase. Less time to test the data.

Neural Networks
Modern deep learning models use artificial neural networks or simply neural networks to extract
information. These neural networks are made up of a simple mathematical function that can be
stacked on top of each other and arranged in the form of layers, giving them a sense of
depth, hence the term Deep Learning.
The neural network is the heart of deep learning models, and it was initially designed to mimic
the working of the neurons in the human brain.
In essence, neural networks enable us to learn the structure of the data or information and help
us to understand it by performing tasks such as clustering, classification, regression, or sample
generation.
Deep learning is implemented with the help of Neural Networks, and the idea behind the
motivation of Neural Network is the biological neurons, which is nothing but a brain cell.
Deep learning is a collection of statistical techniques of machine learning for learning feature
hierarchies that are actually based on artificial neural networks.
So basically, deep learning is implemented by the help of deep networks, which are nothing
but neural networks with multiple hidden layers.

Example of Deep Learning

In the example given above, we provide the raw data of images to the first layer of the input
layer. After then, these input layer will determine the patterns of local contrast that means it
will differentiate on the basis of colors, luminosity, etc. Then the 1st hidden layer will
determine the face feature, i.e., it will fixate on eyes, nose, and lips, etc. And then, it will
fixate those face features on the correct face template. So, in the 2 nd hidden layer, it will
actually determine the correct face here as it can be seen in the above image, after which it
will be sent to the output layer. Likewise, more hidden layers can be added to solve more
complex problems, So, as and when the hidden layers increase, we are able to solve complex
problems.
Architectures:
1. Deep Neural Network – It is a neural network with a certain level of complexity
(having multiple hidden layers in between input and output layers). They are
capable of modeling and processing non-linear relationships.
2. Deep Belief Network (DBN) – It is a class of Deep Neural Network. It is multi-
layer belief networks.
Steps for performing DBN:
a. Learn a layer of features from visible units using Contrastive Divergence
algorithm.
b. Treat activations of previously trained features as visible units and then learn
features of features.
c. Finally, the whole DBN is trained when the learning for the final hidden layer is
achieved.
3. Recurrent (perform same task for every element of a sequence) Neural Network –
Allows for parallel and sequential computation. Similar to the human brain (large
feedback network of connected neurons). They are able to remember important
things about the input they received and hence enables them to be more precise.
Types of Deep Learning Networks

1. Feed Forward Neural Network


A feed-forward neural network is none other than an Artificial Neural Network, which
ensures that the nodes do not form a cycle. In this kind of neural network, all the perceptrons
are organized within layers, such that the input layer takes the input, and the output layer
generates the output. Since the hidden layers do not link with the outside world, it is named
as hidden layers. Each of the perceptrons contained in one single layer is associated with each
node in the subsequent layer. It can be concluded that all of the nodes are fully connected. It
does not contain any visible or invisible connection between the nodes in the same layer.
There are no back-loops in the feed-forward network. To minimize the prediction error, the
backpropagation algorithm can be used to update the weight values.
Applications:
o Data Compression o Speech Recognition
o Pattern Recognition o Handwritten Characters
o Computer Vision Recognition
o Sonar Target Recognition
2. Recurrent Neural Network
Recurrent neural networks are yet another variation of feed-forward networks. Here each of
the neurons present in the hidden layers receives an input with a specific delay in time. The
Recurrent neural network mainly accesses the preceding info of existing iterations. For
example, to guess the succeeding word in any sentence, one must have knowledge about the
words that were previously used. It not only processes the inputs but also shares the length as
well as weights crossways time. It does not let the size of the model to increase with the
increase in the input size. However, the only problem with this recurrent neural network is
that it has slow computational speed as well as it does not contemplate any future input for
the current state. It has a problem with reminiscing prior information.
Applications:
o Machine Translation o Speech Synthesis
o Robot Control o Time Series Anomaly Detection
o Time Series Prediction o Rhythm Learning
o Speech Recognition o Music Composition
3. Convolutional Neural Network
Convolutional Neural Networks are a special kind of neural network mainly used for image
classification, clustering of images and object recognition. DNNs enable unsupervised
construction of hierarchical image representations. To achieve the best accuracy, deep
convolutional neural networks are preferred more than any other neural network.
Applications:
o Identify Faces, Street Signs, o Anomaly Detection.
Tumors. o Drug Discovery.
o Image Recognition. o Checkers Game.
o Video Analysis. o Time Series Forecasting.
o NLP.

4. Restricted Boltzmann Machine


RBMs are yet another variant of Boltzmann Machines. Here the neurons present in the input
layer and the hidden layer encompasses symmetric connections amid them. However, there is
no internal association within the respective layer. But in contrast to RBM, Boltzmann
machines do encompass internal connections inside the hidden layer. These restrictions in
BMs helps the model to train efficiently.
Applications:
o Filtering. o Risk Detection.
o Feature Learning. o Business and Economic analysis.
o Classification.

5. Autoencoders
An autoencoder neural network is another kind of unsupervised machine learning algorithm.
Here the number of hidden cells is merely small than that of the input cells. But the number
of input cells is equivalent to the number of output cells. An autoencoder network is trained
to display the output similar to the fed input to force AEs to find common patterns and
generalize the data. The autoencoders are mainly used for the smaller representation of the
input. It helps in the reconstruction of the original data from compressed data. This algorithm
is comparatively simple as it only necessitates the output identical to the input.
o Encoder: Convert input data in lower dimensions.
o Decoder: Reconstruct the compressed data.
Applications:
o Classification.
o Clustering.
o Feature Compression.

6. Transformers
Transformers are the new class deep learning model that is used mostly for the tasks related to
modeling sequential data, like that in NLP. It is much more powerful than RNNs and they are
replacing them in every task.
How to Work Deep Learning:

First, we need to identify the actual problem


in order to get the right solution and it
should be understood, the feasibility of the
Deep Learning should also be checked
(whether it should fit Deep Learning or not).
Second, we need to identify the relevant
data which should correspond to the actual
problem and should be prepared
accordingly.
Third, Choose the Deep Learning Algorithm
appropriately.
Fourth, Algorithm should be used while
training the dataset.
Fifth, Final testing should be done on the
dataset.

Advantages: 4. Identifies defects easily that are


difficult to detect.
1. Best in-class performance on
problems.
2. Reduces need for feature Disadvantages:
engineering. 1. Large amount of data required.
3. Eliminates unnecessary costs. 2. Computationally expensive to train.
3. No strong theoretical foundation.
Deep Learning Applications:
Deep Learning finds applications in:
1. Automatic Text Generation – Corpus of text is learned and from this model new text is
generated, word-by-word or character-by-character. Then this model is capable of
learning how to spell, punctuate, form sentences, or it may even capture the style.
2. Healthcare – Helps in diagnosing various diseases and treating it.
3. Automatic Machine Translation – Certain words, sentences or phrases in one language
is transformed into another language (Deep Learning is achieving top results in the
areas of text, images).
4. Image Recognition – Recognizes and identifies peoples and objects in images as well as
to understand content and context. This area is already being used in Gaming, Retail,
Tourism, etc.
5. Predicting Earthquakes – Teaches a computer to perform viscoelastic computations
which are used in predicting earthquakes.
6. Speech recognition: Some of the familiar software like Apple’s Siri, Google’s Alexa, and
Microsoft Cortana are based on deep neural networks.
7. Pattern recognition: Pattern recognition is very useful in medical and life sciences.
These algorithms can help radiologists to find tumor cells in the CT scans or even help
them to understand the mechanism behind protein folding. Furthermore, other areas such
as finance can also use pattern recognition systems to detect fraudulent transactions.
8. NLP: Natural language processing or NLP is one of the hot topics in deep learning these
days. Modern architectures like the transformers have revolutionized and improved
machine translation and language modeling. One of such models is GPT3 by Ope; it has
almost reached general intelligence in all NLP tasks.
9. Recommender systems: Recommender systems are on almost every social media
platform these days from Instagram to YouTube and Netflix. These companies use a
recommendation system to recommend shows, videos, posts, and stories based on users'
activities.
10. Self-Driving Cars: In self-driven cars, it is able to capture the images around it by
processing a huge amount of data, and then it will decide which actions should be
incorporated to take a left or right or should it stop. So, accordingly, it will decide what
actions it should take, which will further reduce the accidents that happen every year.
11. Voice Controlled Assistance: When we talk about voice control assistance, then Siri is
the one thing that comes into our mind. So, you can tell Siri whatever you want it to do it
for you, and it will search it for you and display it for you.

Deep Learning Limitations

Data availability
Deep learning models require a lot of data to learn the representation, structure, distribution, and
pattern of the data. If there isn't enough varied data available, then the model will not learn well
and will lack generalization (it won't perform well on unseen data).
The complexity of the model
Designing a deep learning model is often a trial and error process.
A simple model is most likely to underfit, i.e. not able to extract information from the training
set, and a very complex model is most likely to overfit, i.e., not able to generalize well on the test
dataset. Deep learning models will perform well when their complexity is appropriate to the
complexity of the data.
Lacks global generalization
A simple neural network can have thousands to tens of thousands of parameters.
The idea of global generalization is that all the parameters in the model should cohesively
update themselves to reduce the generalization error or test error as much as possible.
However, because of the complexity of the model, it is very difficult to achieve zero
generalization error on the test set.
Incapable of Multitasking
Deep neural networks are incapable of multitasking.
These models can only perform targeted tasks, i.e., process data on which they are trained. For
instance, a model trained on classifying cats and dogs will not classify men and women.
Furthermore, applications that require reasoning or general intelligence are completely beyond
what the current generation’s deep learning techniques can do, even with large sets of data.
Hardware dependence
These models are so complex that a normal CPU will not be able to withstand the computational
complexity. However, multicore high-performing graphics processing units (GPUs) and tensor
processing units (TPUs) are required to effectively train these models in a shorter time.
Big Data
What is Big Data?
Data which are very large in size is called Big Data. Normally we work on data of size
MB(WordDoc ,Excel) or maximum GB(Movies, Codes) but data in Peta bytes i.e. 10^15 byte
size is called Big Data. It is stated that almost 90% of today's data has been generated in the
past 3 years.
Big data is the valuable and powerful fuel that drives large IT industries of the 21st century.
Big data is a spreading technology used in each business sector .

Sources of Big Data


These data come from many sources like
o Social networking sites: Facebook, Google, LinkedIn all these sites generates huge
amount of data on a day to day basis as they have billions of users worldwide.
o E-commerce site: Sites like Amazon, Flipkart, Alibaba generates huge amount of
logs from which users buying trends can be traced.
o Weather Station: All the weather station and satellite gives very huge data which are
stored and manipulated to forecast weather.
o Telecom company: Telecom giants like Airtel, Vodafone study the user trends and
accordingly publish their plans and for this they store the data of its million users.
o Share Market: Stock exchange across the world generates huge amount of data
through its daily transaction.

Big Data Characteristics

Big Data contains a large amount of data that is not being processed by traditional data
storage or the processing unit. It is used by many multinational companies to process the
data and business of many organizations. The data flow would exceed 150 exabytes per day
before replication.
There are five v's of Big Data that explains the characteristics. 5 V's of Big Data
Volume
The name Big Data itself is related to an enormous size. Big Data is a vast 'volumes' of data
generated from many sources daily, such as business processes, machines, social media
platforms, networks, human interactions, and many more.
Facebook can generate approximately a billion messages, 4.5 billion times that the "Like"
button is recorded, and more than 350 million new posts are uploaded each day. Big data
technologies can handle large amounts of data.
Variety
Big Data can be structured, unstructured, and semi-structured that are being collected from
different sources. Data will only be collected from databases and sheets in the past, But these
days the data will comes in array forms, that are PDFs, Emails, audios, SM posts, photos,
videos, etc.

The data is categorized as below:


a. Structured data: In Structured schema, along with all the required columns. It is in a tabular
form. Structured Data is stored in the relational database management system.
b. Semi-structured: In Semi-structured, the schema is not appropriately defined, e.g., JSON,
XML, CSV, TSV, and email. OLTP (Online Transaction Processing) systems are built to work
with semi-structured data. It is stored in relations, i.e., tables.
c. Unstructured Data: All the unstructured files, log files, audio files, and image files are
included in the unstructured data. Some organizations have much data available, but they did not
know how to derive the value of data since the data is raw.
d. Quasi-structured Data:The data format contains textual data with inconsistent data formats that
are formatted with effort and time with some tools.
Veracity
Veracity means how much the data is reliable. It has many ways to filter or translate the data.
Veracity is the process of being able to handle and manage data efficiently. Big Data is also
essential in business development. For example, Facebook posts with hashtags.
Value
Value is an essential characteristic of big data. It is not the data that we process or store. It
is valuable and reliable data that we store, process, and also analyze.
Velocity
Velocity plays an important role compared to others. Velocity creates the speed by which the
data is created in real-time. It contains the linking of incoming data sets speeds, rate of
change, and activity bursts. The primary aspect of Big Data is to provide demanding data
rapidly.
Big data velocity deals with the speed at the data flows from sources like application logs,
business processes, networks, and social media sites, sensors, mobile devices, etc.

Applications of Big Data

The term Big Data is referred to as large amount of complex and unprocessed data. Now a day's
companies use Big Data to make business more informative and allows to take business
decisions by enabling data scientists, analytical modelers and other professionals to analyse large
volume of transactional data.
Travel and Tourism
Travel and tourism are the users of Big Data. It enables us to forecast travel facilities
requirements at multiple locations, improve business through dynamic pricing, and many more.
Financial and banking sector
The financial and banking sectors use big data technology extensively. Big data analytics
help banks and customer behaviour on the basis of investment patterns, shopping trends,
motivation to invest, and inputs that are obtained from personal or financial backgrounds.
Healthcare
Big data has started making a massive difference in the healthcare sector, with the help
of predictive analytics, medical professionals, and health care personnel. It can
produce personalized healthcare and solo patients also.
Telecommunication and media
Telecommunications and the multimedia sector are the main users of Big Data. There
are zettabytes to be generated every day and handling large-scale data that require big data
technologies.
Government and Military
The government and military also used technology at high rates. We see the figures that
the government makes on the record. In the military, a fighter plane requires to
process petabytes of data.
Government agencies use Big Data and run many agencies, managing utilities, dealing with
traffic jams, and the effect of crime like hacking and online fraud.
Aadhar Card: The government has a record of 1.21 billion citizens. This vast data is analyzed
and store to find things like the number of youth in the country
E-commerce
E-commerce is also an application of Big data. It maintains relationships with customers that is
essential for the e-commerce industry. E-commerce websites have many marketing ideas to retail
merchandise customers, manage transactions, and implement better strategies of innovative ideas
to improve businesses with Big data.
o Amazon: Amazon is a tremendous e-commerce website dealing with lots of traffic daily.
But, when there is a pre-announced sale on Amazon, traffic increase rapidly that may
crash the website. So, to handle this type of traffic and data, it uses Big Data. Big Data
help in organizing and analyzing the data for far use.
Social Media
Social Media is the largest data generator. The statistics have shown that around 500+ terabytes
of fresh data generated from social media daily, particularly on Facebook. The data mainly
contains videos, photos, message exchanges, etc. A single activity on the social media site
generates many stored data and gets processed when required. The data stored is in terabytes
(TB); it takes a lot of time for processing. Big Data is a solution to the problem.
How Big Data Works
Big data gives you new insights that open up new opportunities and business models. Getting
started involves three key actions:
1. Integrate
Big data brings together data from many disparate sources and applications. Traditional data
integration mechanisms, such as extract, transform, and load (ETL) generally aren’t up to the
task. It requires new strategies and technologies to analyze big data sets at terabyte, or even
petabyte, scale.
During integration, you need to bring in the data, process it, and make sure it’s formatted and
available in a form that your business analysts can get started with.
2. Manage
Big data requires storage. Your storage solution can be in the cloud, on premises, or both. You
can store your data in any form you want and bring your desired processing requirements and
necessary process engines to those data sets on an on-demand basis. Many people choose their
storage solution according to where their data is currently residing. The cloud is gradually
gaining popularity because it supports your current compute requirements and enables you to
spin up resources as needed.
3. Analyze
Your investment in big data pays off when you analyze and act on your data. Get new clarity
with a visual analysis of your varied data sets. Explore the data further to make new discoveries.
Share your findings with others. Build data models with machine learning and artificial
intelligence. Put your data to work.
Before businesses can put big data to work for them, they should consider how it flows among a
multitude of locations, sources, systems, owners and users. There are five key steps to taking
charge of this "big data fabric" that includes traditional, structured data along with unstructured
and semi-structured data:

I. Set a big data strategy.


II. Identify big data sources.
III. Access, manage and store the data.
IV. Analyze the data.
V. Make intelligent, data-driven decisions.

I) Set a big data strategy


At a high level, a big data strategy is a plan designed to help you oversee and improve the way
you acquire, store, manage, share and use data within and outside of your organization. A big
data strategy sets the stage for business success amid an abundance of data. When developing a
strategy, it’s important to consider existing – and future – business and technology goals and
initiatives. This calls for treating big data like any other valuable business asset rather than just a
byproduct of applications.
II) Identify big data sources
 Streaming data comes from the Internet of Things (IoT) and other connected devices
that flow into IT systems from wearables, smart cars, medical devices, industrial
equipment and more. You can analyze this big data as it arrives, deciding which data to
keep or not keep, and which needs further analysis.
 Social media data stems from interactions on Facebook, YouTube, Instagram, etc. This
includes vast amounts of big data in the form of images, videos, voice, text and sound –
useful for marketing, sales and support functions. This data is often in unstructured or
semistructured forms, so it poses a unique challenge for consumption and analysis.
 Publicly available data comes from massive amounts of open data sources like the US
government’s data.gov, the CIA World Factbook or the European Union Open Data
Portal.
 Other big data may come from data lakes, cloud data sources, suppliers and customers.

III) Access, manage and store big data


Modern computing systems provide the speed, power and flexibility needed to quickly
access massive amounts and types of big data. Along with reliable access, companies also
need methods for integrating the data, building data pipelines, ensuring data quality,
providing data governance and storage, and preparing the data for analysis. Some big data
may be stored on-site in a traditional data warehouse .
IV) Analyze the data
With high-performance technologies like grid computing or in-memory analytics, organizations
can choose to use all their big data for analyses. Big data analytics is how companies gain value
and insights from data. Increasingly, big data feeds today’s advanced analytics endeavors such as
artificial intelligence (AI) and machine learning.
V) Make intelligent, data-driven decisions
Well-managed, trusted data leads to trusted analytics and trusted decisions. To stay competitive,
businesses need to seize the full value of big data and operate in a data-driven way – making
decisions based on the evidence presented by big data rather than gut instinct. The benefits of
being data driven are clear. Data-driven organizations perform better, are operationally more
predictable and are more profitable.
Advantages of Big Data Processing
 Businesses can utilize outside intelligence while taking decisions
 Access to social data from search engines and sites like facebook, twitter are enabling
organizations to fine tune their business strategies.
 Improved customer service
 Traditional customer feedback systems are getting replaced by new systems designed
with Big Data technologies. In these new systems, Big Data and natural language
processing technologies are being used to read and evaluate consumer responses.
 Early identification of risk to the product/services, if any
 Better operational efficiency
Big Data technologies can be used for creating a staging area or landing zone for new data before
identifying what data should be moved to the data warehouse.
Big data challenges
While big data holds a lot of promise, it is not without its challenges.
First, big data is…big. Although new technologies have been developed for data storage, data
volumes are doubling in size about every two years. Organizations still struggle to keep pace
with their data and find ways to effectively store it.
But it’s not enough to just store the data. Data must be used to be valuable and that depends on
curation.
Clean data, or data that’s relevant to the client and organized in a way that enables meaningful
analysis, requires a lot of work. Data scientists spend 50 to 80 percent of their time curating and
preparing data before it can actually be used.
Finally, big data technology is changing at a rapid pace. A few years ago, Apache Hadoop was
the popular technology used to handle big data. Then Apache Spark was introduced in 2014.
Today, a combination of the two frameworks appears to be the best approach. Keeping up with
big data technology is an ongoing challenge.
What is Hadoop

Hadoop is an open source framework from Apache and is used to store process and analyze data
which are very huge in volume. Hadoop is written in Java and is not OLAP (online analytical
processing). It is used for batch/offline processing.It is being used by Facebook, Yahoo, Google,
Twitter, LinkedIn and many more.
Modules of Hadoop
1. HDFS: Hadoop Distributed File System. Google published its paper GFS and on the
basis of that HDFS was developed. It states that the files will be broken into blocks and
stored in nodes over the distributed architecture.
2. Yarn: Yet another Resource Negotiator is used for job scheduling and manage the
cluster.
3. Map Reduce: This is a framework which helps Java programs to do the parallel
computation on data using key value pair. The Map task takes input data and converts it
into a data set which can be computed in Key value pair. The output of Map task is
consumed by reduce task and then the out of reducer gives the desired result.
4. Hadoop Common: These Java libraries are used to start Hadoop and are used by other
Hadoop modules.

Pervasive Computing
Pervasive Computing is also called as Ubiquitous computing, and it is the new trend toward
embedding everyday objects with microprocessors so that they can communicate information. It
refers to the presence of computers in common objects found all around us so that people are
unaware of their presence. All these devices communicate with each other over wireless
networks without the interaction of the user.
It is also known as ubiquitous computing. The terms ubiquitous and pervasive signify "existing
everywhere." Pervasive computing systems are totally connected and consistently available.
Pervasive computing, is the growing trend of embedding computational capability (generally in
the form of microprocessors) into everyday objects to make them effectively communicate and
perform useful tasks in a way that minimizes the end user's need to interact with computers as
computers. Pervasive computing devices are network-connected and constantly available.
Unlike desktop computing, pervasive computing can occur with any device, at any time, in any
place and in any data format across any network and can hand tasks from one computer to
another.
Pervasive computing devices have evolved to include:
 laptops;
 notebooks;
 smartphones;
 tablets;
 wearable devices and
 sensors (for example, on fleet management, lighting systems, appliances).
How ubiquitous computing is used?
Pervasive computing applications have been designed for consumer use and to help people do
their jobs.
An example of pervasive computing is an Apple Watch that alerts the user to a phone call and
allows the call to be completed through the watch. Another example is when a registered user for
Audible, Amazon's audio book server, starts his or her book using the Audible app on a
smartphone on the train and continues listening to the book through Amazon Echo at home.
An environment in which devices, present everywhere, are capable of some form of computing
can be considered a ubiquitous computing environment.
Pervasive computing is a combination of three technologies, namely:
1. Micro electronic technology:
This technology gives small powerful device and display with low energy consumption.
2. Digital communication technology:
This technology provides higher bandwidth, higher data transfer rate at lower costs and
with world wide roaming.
3. The Internet standardization:
This standardization is done through various standardization bodies and industry to give
the framework for combining all components into an interoperable system with
security, service and billing systems.
Thus, wireless communication, consumer electronics and computer technology were all merged
into one to create a new environment called pervasive computing environment. It helps to access
information and render modern administration in areas that do not have a traditional wire-based
computing environment.
Pervasive computing is the next dimension of personal computing in the near future, and it will
definitely change and improve our work environment and communication methods.
Key Characteristics of Pervasive computing:

1. Many devices can be integrated into one system for multi-purpose uses.
2. A huge number of various interfaces can be used to build an optimized user interface.
3. Concurrent operation of online and offline supported.
4. A large number of specialized computers are integrated through local buses and Internet.
5. Security elements are added to prevent misuse and unauthorized access.
6. Personalization of functions adapts the systems to the user’s preferences, so that no PC
knowledge is required of the user to use and manage the system.
Examples
Examples of pervasive computing include electronic toll systems on highways; tracking
applications, such as Life360, which can track the location of the user, the speed at which they
are driving and how much battery life their smartphone has; Apple Watch; Amazon Echo; smart
traffic lights; and Fitbit.
Applications:
There are a rising number of pervasive devices available in the market nowadays. The areas of
application of these devices include:

 Retail  Tracking
 Airlines booking and check-in  Car information System
 Sales force automation  Email access via WAP (Wireless
 Healthcare Application Protocol) and voice.

For example, in retail industry, there is a requirement for faster and cheaper methods to bring
goods to the consumer from stores via Internet. Mobile computers are provided with bar code
readers for tracking the product during manufacture. Currently consumers use computers to
select products. In future, they will use PDA (Personal Digital Assistant) and pervasive devices
in the domestic markets too. When they complete writing the list of items to be bought on these
devices, this list can then be sent to the supermarket, and purchase can be delivered to the
consumer. The advantages of this are faster processing of data and execution of data mining.
Importance
Pervasive computing systems are capable of collecting, processing and communicating data, they
can adapt to the data's context and activity. That means a network that can understand its
surroundings and improve the human experience and quality of life.
Advantages of pervasive computing
As described above, pervasive computing requires less human interaction than a ubiquitous
computing environment where there may be more connected devices, but that the extraction and
processing of data requires more intervention.
Because pervasive computing systems are capable of collecting, processing and communicating
data, they can adapt to the data's context and activity. That means, in essence, that a network that
can understand its surroundings and improve the human experience and quality of life.
Disadvantages of pervasive computing
A distinct problem with pervasive computing is that it is not entirely secure. The devices and
technologies used in pervasive computing do not lend themselves well to typical data security.
Pervasive computing include frequent line connections that are broken, slow connections, very
expensive operating costs, host bandwidths that are limited in nature and location-dependent
data.
All of these instances can impede the security of pervasive computing because they result in
multiple system vulnerabilities.
Cloud Computing
What is Cloud Computing?
Cloud Computing can be defined as
delivering computing power (CPU, RAM,
Network Speeds, Storage OS software) a
service over a network (usually on the
internet) rather than physically having the
computing resources at the customer
location.

Example: AWS, Azure, Google Cloud


Let’s learn Cloud computing with an example -
Whenever you travel through a bus or train, you take a ticket for your destination and hold back
to your seat till you reach your destination. Likewise other passengers also takes ticket and travel
in the same bus with you and it hardly bothers you where they go. When your stop comes you
get off the bus thanking the driver. Cloud computing is just like that bus, carrying data and
information for different users and allows to use its service with minimal cost.
Why the Name Cloud?

The term “Cloud” came from a network


design that was used by network engineers
to represent the location of various network
devices and there inter-connection. The
shape of this network design was like a
cloud.

Why Cloud Computing?


With increase in computer and Mobile user’s, data storage has become a priority in all fields.
Large and small scale businesses today thrive on their data & they spent a huge amount of
money to maintain this data. It requires a strong IT support and a storage hub. Not all businesses
can afford high cost of in-house IT infrastructure and back up support services. For them Cloud
Computing is a cheaper solution. Perhaps its efficiency in storing data, computation and less
maintenance cost has succeeded to attract even bigger businesses as well.
Cloud computing decreases the hardware and software demand from the user’s side. The only
thing that user must be able to run is the cloud computing systems interface software, which can
be as simple as Web browser, and the Cloud network takes care of the rest.
While accessing e-mail service our data is stored on cloud server and not on our computer. The
technology and infrastructure behind the cloud is invisible.
Benefits of Cloud Computing
The potential for cost saving is the major reason of cloud services adoption by many
organizations. Cloud computing gives the freedom to use services as per the requirement and
pay only for what you use. Due to cloud computing it has become possible to run IT operations
as a outsourced unit without much in-house resources.
Following are the benefits of cloud computing:
1. Lower IT infrastructure and 5. Improved compatibility between
computer costs for users Operating systems
2. Improved performance 6. Backup and recovery
3. Fewer Maintenance issues 7. Performance and Scalability
4. Instant software updates 8. Increased storage capacity
9. Increase data safety
Types of Clouds
There are four different cloud models that
you can subscribe according to business
needs:

1. Private Cloud: Here, computing resources are deployed for one particular organization.
This method is more used for intra-business interactions. Where the computing resources
can be governed, owned and operated by the same organization.
2. Community Cloud: Here, computing resources are provided for a community and
organizations.
3. Public Cloud: This type of cloud is used usually for B2C (Business to Consumer) type
interactions. Here the computing resource is owned, governed and operated by
government, an academic or business organization.
4. Hybrid Cloud: This type of cloud can be used for both type of interactions - B2B
(Business to Business) or B2C ( Business to Consumer). This deployment method is
called hybrid cloud as the computing resources are bound together by different clouds.
Cloud Computing Services
The three major Cloud Computing Offerings are

 Software as a Service (SaaS)


 Platform as a Service (PaaS)
 Infrastructure as a Service (IaaS)
Different business use some or all of these components according to their requirement.

SaaS (Software as a Service)


SaaS or software as a service is a software distribution model in which applications are
hosted by a vendor or service provider and made available to customers over a network
(internet). SaaS is becoming an increasingly prevalent delivery model as underlying
technologies that supports Service Oriented Architecture (SOA) or Web Services.
Through internet this service is available to users anywhere in the world.
Traditionally, software application needed to be purchased upfront &then installed it onto
your computer. SaaS users on the other hand, instead of purchasing the software subscribes to
it, usually on monthly basis via internet.
Anyone who needs an access to a particular piece of software can be subscribe as a user,
whether it is one or two people or every thousands of employees in a corporation. SaaS is
compatible with all internet enabled devices.
Many important tasks like accounting, sales, invoicing and planning all can be performed
using SaaS.

PaaS (Platform as a Service)


Platform as a service, is referred as PaaS, it provides a platform and environment to allow
developers to build applications and services. This service is hosted in the cloud and accessed
by the users via internet.
To understand in a simple terms, let compare this with painting a picture, where you are
provided with paint colors, different paint brushes and paper by your school teacher and you
just have to draw a beautiful picture using those tools.
PaaS services are constantly updated & new features added. Software developers, web
developers and business can benefit from PaaS. It provides platform to support application
development. It includes software support and management services, storage, networking,
deploying, testing, collaborating, hosting and maintaining applications.

IaaS (Infrastructure as a Service)


IaaS (Infrastructure As A Service) is one of the fundamental service model of cloud
computing alongside PaaS ( Platform as a Service). It provides access to computing resources
in a virtualized environment “the cloud” on internet. It provides computing infrastructure
like virtual server space, network connections, bandwidth, load balancers and IP addresses.
The pool of hardware resource is extracted from multiple servers and networks usually
distributed across numerous data centers. This provides redundancy and reliability to IaaS.
IaaS (Infrastructure as a service) is a complete package for computing. For small scale
businesses who are looking for cutting cost on IT infrastructure, IaaS is one of the solutions.
Annually a lot of money is spent in maintenance and buying new components like hard-
drives, network connections, external storage device etc. which a business owner could have
saved for other expenses by using IaaS.
What is Cloud Computing Architecture?
Let’s have a look into Cloud Computing and see what Cloud Computing is made of. Cloud
computing comprises of two components front end and back end. Front end consist client
part of cloud computing system. It comprise of interfaces and applications that are required to
access the cloud computing platform.

While back end refers to the cloud itself, it comprises of the resources that are required for
cloud computing services. It consists of virtual machines, servers, data storage, security
mechanism etc. It is under providers control.
Cloud computing distributes the file system that spreads over multiple hard disks and
machines. Data is never stored in one place only and in case one unit fails the other will take
over automatically. The user disk space is allocated on the distributed file system, while
another important component is algorithm for resource allocation. Cloud computing is a
strong distributed environment and it heavily depends upon strong algorithm.
Security concerns for Cloud Computing
While using cloud computing, the major issue that concerns the users is about its security.
One concern is that cloud providers themselves may have access to customer’s unencrypted
data- whether it’s on disk, in memory or transmitted over the network.
To provide security for systems, networks and data cloud computing service providers have
joined hands with TCG (Trusted Computing Group) which is non-profit organization which
regularly releases a set of specifications to secure hardware, create self-encrypting drives and
improve network security. It protects the data from root kits and malware.
As computing has expanded to different devices like hard disk drives and mobile phones,
TCG has extended the security measures to include these devices. It provides ability to create
a unified data protection policy across all clouds.
Some of the trusted cloud services are Amazon, Box.net, Gmail and many others.
Privacy Concern & Cloud Computing
Privacy present a strong barrier for users to adapt into Cloud Computing systems
There are certain measures which can improve privacy in cloud computing.
1. The administrative staff of the cloud computing service could theoretically monitor
the data moving in memory before it is stored in disk. To keep the confidentiality of a
data, administrative and legal controls should prevent this from happening.
2. The other way for increasing the privacy is to keep the data encrypted at the cloud
storage site, preventing unauthorized access through the internet; even cloud vendor
can’t access the data either.
Artificial Intelligence
Artificial Intelligence (AI), or machine intelligence, is the field developing computers and
robots capable of parsing data contextually to provide requested information, supply analysis,
or trigger events based on findings. Through techniques like machine learning and neural
networks, companies globally are investing in teaching machines to ‘think’ more like
humans.
What is Artificial Intelligence (AI)?
Artificial Intelligence, or simply AI, is the term used to describe a machine’s ability to
simulate human intelligence. Actions like learning, logic, reasoning, perception, creativity,
that were once considered unique to humans, is now being replicated by technology and used
in every industry.
A common example of AI in today’s world is chatbots, specifically the “live chat” versions
that handle basic customer service requests on company websites.
How does AI work?
Artificial Intelligence is a complex field with many components and methodologies used to
achieve the final result — an intelligent machine. AI was developed by studying the way the
human brain thinks, learns and decides, then applying those biological mechanisms to
computers.
As opposed to classical computing, where coders provide the exact inputs, outputs, and logic,
artificial intelligence is based on providing a machine the inputs and a desired outcome,
letting the machine develop its own path to achieve its set goal. This frequently allows
computers to better optimize a situation than humans, such as optimizing supply chain
logistics and streamlining financial processes.
Types of AI
There are four types of AI that differ in their complexity of abilities:
1. Reactive machines use previous data to draw conclusions for current decisions, such
as a chess AI.
2. Limited memory uses live data to read a situation and make decisions, such as self-
driving cars.
3. Theory of mind is something we haven’t reached yet, and deals with AI understanding
that every entity has its own set of underlying concepts — such as motives, intentions
and emotions.
4. Self-awareness is the final stage of AI, where an AI can not only understand others’
consciousness (theory of mind) but also has a concept of its own existence.
Examples of AI
Artificial intelligence is used in virtually all businesses; in fact, you likely interact with it in
some capacity on a daily basis. Chatbots, smart cars, IoT devices, healthcare, banking,
and logistics all use artificial intelligence to provide a superior experience.
One AI that is quickly finding its way into most consumer’s homes is the voice assistant,
such as Apple’s Siri, Amazon’s Alexa, Google’s Assistant, and Microsoft’s Cortana.
Once simply considered part of a smart speaker, AI-equipped voice assistants are now
powerful tools deeply integrated across entire ecosystems of channels and devices to provide
an almost human-like virtual assistant experience.
Benefits of AI
Artificial intelligence can help reduce human error, create more precise analytics, and turn
data collecting devices into powerful diagnostic tools.
One example of this is wearable devices such as smartwatches and fitness trackers, which put
data in the hands of consumers to empower them to play a more active role managing their
health.

Advancements in AI
In the last five years, the field of AI has made major progress in almost all its standard sub-areas,
including vision, speech recognition and generation, natural language processing (understanding
and generation), image and video generation, multi-agent systems, planning, decision-making, and
integration of vision and motor control for robotics.
In addition, breakthrough applications emerged in a variety of domains including games, medical
diagnosis, logistics systems, autonomous driving, language translation, and interactive personal
assistance.
Language Processing
Language processing technology made a major leap in the last five years, leading to the
development of network architectures with enhanced capability to learn from complex and
context-sensitive data. These models’ facility with language is already supporting
applications such as machine translation, text classification, speech recognition, writing aids,
and chatbots. Future applications could include improving human-AI interactions across
diverse languages and situations.
Computer Vision and Image Processing
Image-processing technology is now widespread, finding uses ranging from video-conference
backgrounds to the photo-realistic images known as deepfakes. Many image-processing
approaches use deep learning for recognition, classification, conversion, and other tasks.
Training time for image processing has been substantially reduced.
Real-time object-detection systems such as YOLO (You Only Look Once) that notice
important objects when they appear in an image are widely used for video surveillance of
crowds and are important for mobile robots including self-driving cars. Face-recognition
technology has also improved significantly over the last five years, and now some
smartphones and even office buildings rely on it to control access
Games
Developing algorithms for games and simulations in adversarial situations has long been a
fertile training ground and a showcase for the advancement of AI techniques.
Robotics
The last five years have seen consistent progress in intelligent robotics driven by machine
learning, powerful computing and communication capabilities, and increased availability of
sophisticated sensor systems.
Mobility
Autonomous vehicles or self-driving cars have been one of the hottest areas in deployed
robotics, as they impact the entire automobile industry as well as city planning. The design of
self-driving cars requires integration of a range of technologies including sensor fusion, AI
planning and decision-making, vehicle dynamics prediction, on-the-fly rerouting, inter-
vehicle communication, and more.
Driver assist systems are increasingly widespread in production vehicles. These systems use
sensors and AI-based analysis to carry out tasks such as adaptive cruise control to safely
adjust speed, and lane-keeping assistance to keep vehicles centered on the road.
Healthcare:
AI is increasingly being used in biomedical applications, particularly in diagnosis, drug
discovery, and basic life science research. Diseases are more quickly and accurately
diagnosed, drug discovery is sped up and streamlined, virtual nursing assistants monitor
patients and big data analysis helps to create a more personalized patient experience. Tools
now exist for identifying a variety of eye and skin disorders, detecting cancers and supporting
measurements needed for clinical diagnosis.
Finance
AI has been increasingly adopted into finance. Deep learning models now partially automate
lending decisions for several lenders and have transformed payments with credit scoring.
These new systems often take advantage of consumer data that are not traditionally used in
credit scoring.
In the space of personal finance, so-called robo-advising—automated financial advice—is
quickly becoming mainstream for investment and overall financial planning. For financial
institutions, uses of AI are going beyond detecting fraud and enhancing cybersecurity to
automating legal and compliance documentation as well as detecting money laundering.
Recommender Systems
With the explosion of information available to us, recommender systems that automatically
prioritize what we see when we are online have become absolutely essential. Such systems
have always drawn heavily on AI, and now they have a dramatic influence on people’s
consumption of products, services, and content—from news, to music, to videos, and more
Transportation: Although it could take a decade or more to perfect them, autonomous cars
will one day ferry us from place to place
Manufacturing: AI powered robots work alongside humans to perform a limited range of
tasks like assembly and stacking, and predictive analysis sensors keep equipment running
smoothly
Education: Textbooks are digitized with the help of AI, early-stage virtual tutors assist
human instructors and facial analysis gauges the emotions of students to help determine
who’s struggling or bored and better tailor the experience to their individual needs
Media: Journalism is harnessing AI, too, and will continue to benefit from it. Bloomberg
uses Cyborg technology to help make quick sense of complex financial reports. The
Associated Press employs the natural language abilities of Automated Insights to produce
3,700 earning reports stories per year — nearly four times more than in the recent past.
Customer Service: Last but hardly least, Google is working on an AI assistant that can place
human-like calls to make appointments at, say, your neighborhood hair salon. In addition to
words, the system understands context and nuance.
IoT (Internet of Things)
IoT stands for Internet of Things, which means accessing and controlling daily usable
equipments and devices using Internet. Internet of Things (IoT) is how we describe the
digitally connected universe of everyday physical devices. These devices are embedded with
internet connectivity, sensors and other hardware that allow communication and control via
the web.
Definition: Connecting everyday things embedded with electronics, software, and sensors to
internet enabling to collect and exchange data without human interaction called as the
Internet of Things (IoT).
The term "Things" in the Internet of Things refers to anything and everything in day to day
life which is accessed or connected through the internet.

IoT is an advanced automation and analytics system which deals with artificial intelligence,
sensor, networking, electronic, cloud messaging etc. to deliver complete systems for the
product or services. The system created by IoT has greater transparency, control and
performance.
For example, a house, where we can connect our home appliances such as air conditioner,
light, etc. through each other and all these things are managed at the same platform. Since we
have a platform, we can connect our car, track its fuel meter, speed level, and also track the
location of the car.

If there is a common platform where all these things can connect to each other would be great
because based on my preference, I can set the room temperature. For example, if I love the
room temperature to to be set at 25 or 26-degree Celsius when I reach back home from my
office, then according to my car location, my AC would start before 10 minutes I arrive at
home. This can be done through the Internet of Things (IoT).
IoT makes once "dumb" devices "smarter" by giving them the ability to send data over the
internet, allowing the device to communicate with people and other IoT-enabled things.
The connected "smart home" is a good example of IoT in action. Internet-enabled
thermostats, doorbells, smoke detectors and security alarms create a connected hub where
data is shared between physical devices and users can remotely control the "things" in that
hub (i.e., adjusting temperature settings, unlocking doors, etc.) via a mobile app or website.
How does Internet of Thing (IoT) Work?
IoT works in the following way:
 Devices have hardware like sensors, for example, that collect data.
 The data collected by the sensors is then shared via the cloud and integrated with
software.
 The software then analyzes and transmits the data to users via an app or website.
The working of IoT is different for different IoT echo system (architecture). However, the
key concept of their working are similar. The entire working process of IoT starts with the
device themselves, such as smartphones, digital watches, electronic appliances, which
securely communicate with the IoT platform. The platforms collect and analyze the data from
all multiple devices and platforms and transfer the most valuable data with applications to
devices.

Features of IoT
The most important features of IoT on which it works are connectivity, analyzing,
integrating, active engagement, and many more. Some of them are listed below:
Connectivity: Connectivity refers to establish a proper connection between all the things of
IoT to IoT platform it may be server or cloud. After connecting the IoT devices, it needs a
high speed messaging between the devices and cloud to enable reliable, secure and bi-
directional communication.
Analyzing: After connecting all the relevant things, it comes to real-time analyzing the data
collected and use them to build effective business intelligence. If we have a good insight into
data gathered from all these things, then we call our system has a smart system.
Integrating: IoT integrating the various models to improve the user experience as well.
Artificial Intelligence: IoT makes things smart and enhances life through the use of data. For
example, if we have a coffee machine whose beans have going to end, then the coffee
machine itself order the coffee beans of your choice from the retailer.
Sensing: The sensor devices used in IoT technologies detect and measure any change in the
environment and report on their status. IoT technology brings passive networks to active
networks. Without sensors, there could not hold an effective or true IoT environment.
Active Engagement: IoT makes the connected technology, product, or services to active
engagement between each other.
Endpoint Management: It is important to be the endpoint management of all the IoT system
otherwise, it makes the complete failure of the system. For example, if a coffee machine itself
order the coffee beans when it goes to end but what happens when it orders the beans from a
retailer and we are not present at home for a few days, it leads to the failure of the IoT
system. So, there must be a need for endpoint management.

IoT - Platform
As in IoT, all the IoT devices are connected to other IoT devices and application to transmit
and receive information using protocols. There is a gap between the IoT device and IoT
application. An IoT Platform fills the gap between the devices (sensors) and application
(network). Thus we can say that an IoT platform is an integrated service that fulfills the
gap between the IoT device and application and offers you to bring physical object
online.

IoT Architecture
There is not such a unique or standard consensus on the Internet of Things (IoT) architecture
which is universally defined. The IoT architecture differs from their functional area and their
solutions. However, the IoT architecture technology mainly consists of four major
components:
Components of IoT Architecture
o Sensors/Devices o Cloud/Management Service Layer
o Gateways and Networks o Application Layer
Stages of IoT Solutions Architecture
The IoT architecture is a fundamental way to design the various elements of IoT, so that it
can deliver services over the networks and serve the needs for the future.
Following are the primary stages (layers) of IoT that provides the solution for IoT
architecture.
1. Sensors/Actuators: Sensors or Actuators are the devices that are able to emit, accept and
process data over the network. These sensors or actuators may be connected either through
wired or wireless. This contains GPS, Electrochemical, Gyroscope, RFID, etc. Most of the
sensors need connectivity through sensors gateways. The connection of sensors or actuators
can be through a Local Area Network (LAN) or Personal Area Network.
2. Gateways and Data Acquisition: As the large numbers of data are produced by this sensors
and actuators need the high-speed Gateways and Networks to transfer the data. This network
can be of type Local Area Network (LAN such as WiFi, Ethernet, etc.), Wide Area Network
(WAN such as GSM, 5G, etc.).
3. Edge IT: Edge in the IoT Architecture is the hardware and software gateways that analyze
and pre-process the data before transferring it to the cloud. If the data read from the sensors
and gateways are not changed from its previous reading value then it does not transfer over
the cloud, this saves the data used.
4. Data center/ Cloud: The Data Center or Cloud comes under the Management Services which
process the information through analytics, management of device and security controls.
Beside this security controls and device management the cloud transfer the data to the end
users application such as Retail, Healthcare, Emergency, Environment, and Energy, etc.
IoT Devices
Internet-of-Things devices are all around us; constantly transmitting data and “talking” with
other IoT devices. We come across IoT devices almost everyday in the form of virtual
assistants, “smart” electronics and in our wearable health trackers. Each Internet-of-Things
device tracks information in real-time and relays it to us to help make our lives safer,
healthier or more efficient.

IoT devices are enlarging the internet connectivity beyond standard devices such as
smartphones, laptops, tablets, and desktops. Embedding these devices with technology enable
us to communicate and interact over the networks and they can be remotely monitored and
controlled.
IoT devices include computer devices, software, wireless sensors, and actuators. These IoT
devices are connected over the internet and enabling the data transfer among objects or
people automatically without human intervention.
Some of the common and popular IoT devices are given below:
Properties of IoT Devices
Some of the essential properties of IoT devices are mention below:
o Sense: The devices that sense its surrounding environment in the form of temperature,
movement, and appearance of things, etc.
o Send and receive data: IoT devices are able to send and receive the data over the
network connection.
o Analyze: The devices can able to analyze the data that received from the other device
over the internet networks.
o Controlled: IoT devices may control from some endpoint also. Otherwise, the IoT
devices are themselves communicate with each other endlessly leads to the system
failure.

Advantages of IoT
Internet of things facilitates the several advantages in day-to-day life in the business sector.
Some of its benefits are given below:
o Efficient resource utilization: If we know the functionality and the way that how each
device work we definitely increase the efficient resource utilization as well as monitor natural
resources.
o Minimize human effort: As the devices of IoT interact and communicate with each other
and do lot of task for us, then they minimize the human effort.
o Save time: As it reduces the human effort then it definitely saves out time. Time is the
primary factor which can save through IoT platform.
o Enhance Data Collection:
o Improve security: Now, if we have a system that all these things are interconnected then we
can make the system more secure and efficient.

Disadvantages of IoT
As the Internet of things facilitates a set of benefits, it also creates a significant set of
challenges. Some of the IoT challenges are given below:
o Security: As the IoT systems are interconnected and communicate over networks. The
system offers little control despite any security measures, and it can be lead the various kinds
of network attacks.
o Privacy: Even without the active participation on the user, the IoT system provides
substantial personal data in maximum detail.
o Complexity: The designing, developing, and maintaining and enabling the large technology
to IoT system is quite complicated.
Advancements and Applications in the IoT
The IoT, along with artificial intelligence, machine learning and cloud technology, has been
one of the most important trends in high-tech over the past couple years. It has been
developing at astonishing speeds since its inception, often rapidly changing direction and
popping up in new and quite unexpected forms.
The rise of 5G technology - 5G networks are at the forefront of development of cellular
mobile communications. Their extremely high speeds will offer an array of new possibilities
for the IoT, paving the way for a degree of connectivity that is impossible with current
standards. Through 5G, data can be gathered, analyzed and managed in real time.
Edge Computing - Edge computing is literally the opposite of cloud computing. Edge
computing means that data is stored in micro-centers as opposed to the cloud,
providing numerous new options for the IoT. By storing data locally, it offers a cheaper,
faster, and more efficient approach to data processing.
Smart Stores - In 2019, smart lighting devices, video feeds and Wi-Fi enabled food
traffic-monitoring software allows store owners to collect info about customer traffic
patterns in the shop, how much time they spend in each of its aisles and how they
interact with products on display. After analyzing this data, retailers are able to change
the way they lay out their merchandise and decide how much of it they put on display or
even change their entire store layouts to enhance them in line with knowledge they
have gained about customer behavior.
Smart Cities - Using the IoT to track the condition of its traffic lights. If one malfunctions,
the system notifies utility companies so that they can quickly send a technician to solve
the problem.
Besides traffic lights, the IoT is being put to use in creating smart common areas in cities
around the world. Sidewalk Labs, owned by Alphabet, the parent of Google, is building a
smart neighborhood in Toronto. Smart sensors are being installed around the neighborhood
that will record everything from shared car use, building occupancy, sewage flow, and
optimum temperature choices around the clock. The goal is to make the neighborhood as
safe, convenient, and comfortable as possible for those who live there. Once this model is
perfected, it could set the standard for other smart neighborhoods and even eventually entire
smart cities.
Manufacturing and Healthcare - The IoT has already begun to transform manufacturing.
Sensors, RFID tags, and smart beacons have been in place for several years. Factory
owners are becoming able to prevent delays, improve production output, reduce
equipment downtime, and of course manage inventory.
In the world of healthcare, already more than half of organizations have adopted IoT
technology. It is an area where there are almost endless possibilities- smart pills, smart home
care, electronic health records, and personal healthcare management.
Connected Smart Cars – This is in the form of diagnostic information about the car.
Everything from tire pressure, oil level, fuel consumption , and when something goes wrong
with the engine is now available to be sent to the palm of your hand via a Wi-Fi connection to
your smartphone.
we can also see more IoT advancement like Connected apps, voice search, and current traffic
information in our cars . While self-driving cars are utilize IoT technology in the multitude
of sensors they contain that allows them to be monitored remotely as they navigate streets.
Increased Security Concerns - In the past, a malware infection meant just lost or
compromised data. The emergence of the IoT means that a virus or ransomware infection can
easily disable vital functions and services, and security vendors are starting to offer endpoint
security solutions to their existing services to prevent data loss and give suggestions about
threat protection and network health.
Healthcare and IoT the perfect combination - Sensors and wearable devices can collect
and monitor the data and instantaneously send it for processing. Health monitors transfer
analytics by the minute which can help doctors across the world analyze and deal with ‘the’
problems.
The general population is becoming more tech savvy which is a necessity for these adoptions.
Mobile applications as well as virtual assistants are becoming more common and widely used
in everyday routines. We even have smart cars coming up which can store and transmit your
medical information on the go. The possibilities are endless.
IoT cloud development - Cloud services have become absolutely essential for IoT
deployment. Because of this, major companies like Amazon are putting in rigorous
efforts to add more offerings to their arsenal catering cloud services.
A couple of years back, Microsoft launched its “Azure IoT Edge” which now allows various
devices in an environment to run cloud services without being actually connected to a cloud.
With cloud services, security features always become a priority thus tech giants are putting in
increased efforts to improve the security layers and locks surrounding these cloud data
storage features.
The prominence of big data and AI - It is a given that with the huge amounts of data
being collected, its processing is a must. Analyzing all this data is important as raw data
in itself cannot be of much use. Thus Big Data management is becoming essential and
growing exponentially with these advancements in IoT. The application of all this data
in real life situations holds great potential.
AI technologies are being employed aggressively by various companies in response to the
rising competition. A strong relationship is being formed therefore between AI, Big Data and
IoT devices and how these entities operate. AI helps in handling and sorting through the wide
aspects of Big Data and marks considerable growth in its proper use.
Introduction of IoT operating systems - We have all been using Windows and iOS for
some time now. But with the way IoT is developing, these operating systems are
becoming insufficient. Real-time responses are required for IoT to function is a timely
manner and many of our conventional operating systems lack that capability.
IoT technologies are also introducing newer and advanced chipsets which might not be
compatible with the current operating systems. Keeping these factors in mind, IoT facilitating
operating systems are being introduced which are specifically designed with keeping IoT
connectivity in mind.
Implementation of IoT into Marketing - Marketing companies all around the world are
making great use of IoT developments and Big Data to target their clients. Advanced
analytics make it possible to target niches in better and effective ways.
Examples of IoT Platforms
There are several IoT Platforms available that provides facility to deploy IoT application
actively. Some of them are listed below:
Amazon Web Services (AWS) IoT platform: Amazon Web Service IoT platform offers a
set of services that connect to several devices and maintain the security as well. This platform
collects data from connected devices and performs real-time actions.
Microsoft Azure IoT platform: Microsoft Azure IoT platform offers strong security
mechanism, scalability and easy integration with systems. It uses standard protocols that
support bi-directional communication between connected devices and platform. Azure IoT
platform has an Azure Stream Analytics that processes a large amount of information in real-
time generated by sensors. Some common features provided by this platform are:
o Information monitoring
o A rules engine
o Device shadowing
o Identity registry
Google Cloud Platform IoT: Google Cloud Platform is a global cloud platform that
provides a solution for IoT devices and applications. It handles a large amount of data using
Cloud IoT Core by connecting various devices. It allows to apply BigQuery analysis or to
apply Machine learning on this data. Some of the features provided by Google Cloud IoT
Platform are:
o Cloud IoT Core
o Speed up IoT devices
o Cloud publisher-subscriber
o Cloud Machine Learning Engine

IBM Watson IoT platform: The IBM Watson IoT platform enables the developer to deploy
the application and building IoT solutions quickly. This platform provides the following
services:
o Real-time data exchange
o Device management
o Secure Communication
o Data sensor and weather data services

Artik Cloud IoT platform: Arthik cloud IoT platform is developed by Samsung to enable
devices to connect to cloud services. It has a set of services that continuously connect devices
to the cloud and start gathering data. It stores the incoming data from connected devices and
combines this information. This platform contains a set of connectors that connect to third-
party services.
Bosch IoT Suite:
Bosch cloud IoT Suit is based on Germany. It offers safe and reliable storing of data on its
server in Germany. This platform supports full app development from prototype to
application development.
How IoT platform help:
o IoT Platform connects sensors and devices.
o IoT platform handles different software communication protocol and hardware.
o IoT platform provides security and authentication for sensors and users.
o It collects, visualizes, and analyzes the data gathered by the sensor and device.
Block chain is a constantly growing ledger that keeps a permanent record of all the
transactions that have taken place in a secure, chronological, and immutable way. It can be
used for the secure transfer of money, property, contracts, etc. without requiring a third-party
intermediary such as bank or government. Block chain is a software protocol, but it could not
be run without the Internet.

What is Block chain?


A block chain is a constantly growing ledger which keeps a permanent record of all the
transactions that have taken place in a secure, chronological, and immutable way.
Let's breakdown the definition,
o Ledger: It is a file that is constantly growing.
o Permanent: It means once the transaction goes inside a blockchain, you can put up it
permanently in the ledger.
o Secure: Blockchain placed information in a secure way. It uses very advanced
cryptography to make sure that the information is locked inside the blockchain.
o Chronological: Chronological means every transaction happens after the previous
one.
o Immutable: It means as you build all the transaction onto the blockchain, this ledger
can never be changed.
A blockchain is a chain of blocks which contain information. Each block records all of the
recent transactions, and once completed goes into the blockchain as a permanent database.
Each time a block gets completed, a new block is generated.

Who uses the blockchain?


Blockchain technology can be integrated into multiple areas. The primary use of blockchains
is as a distributed ledger for cryptocurrencies. It shows great promise across a wide range of
business applications like Banking, Finance, Government, Healthcare, Insurance, Media and
Entertainment, Retail, etc.
Need of Blockchain

Blockchain technology has become popular because of the following.


o Time reduction: In the financial industry, blockchain can allow the quicker
settlement of trades. It does not take a lengthy process for verification, settlement, and
clearance. It is because of a single version of agreed-upon data available between all
stakeholders.
o Unchangeable transactions: Blockchain register transactions in a chronological
order which certifies the unalterability of all operations, means when a new block is
added to the chain of ledgers, it cannot be removed or modified.
o Reliability: Blockchain certifies and verifies the identities of each interested parties.
This removes double records, reducing rates and accelerates transactions.
o Security: Blockchain uses very advanced cryptography to make sure that the
information is locked inside the blockchain. It uses Distributed Ledger Technology
where each party holds a copy of the original chain, so the system remains operative,
even the large number of other nodes fall.
o Collaboration: It allows each party to transact directly with each other without
requiring a third-party intermediary.
o Decentralized: It is decentralized because there is no central authority supervising
anything. There are standards rules on how every node exchanges the blockchain
information. This method ensures that all transactions are validated, and all valid
transactions are added one by one.

Blockchain Version

The evolution of blockchain technology and its versioning from 1.0 to 3.0 are explained
below.
Blockchain 1.0: Currency
The idea of creating money through solving computational puzzles was first introduced
in 2005 by Hal Finney, who created the first concept for cryptocurrencies (The
implementation of distributed ledger technology). This ledger allows financial transactions
based on blockchain technology or DLT to be executed with Bitcoin. Bitcoin is the most
prominent example in this segment. It is being used as cash for the Internet and seen as the
enabler of an Internet of Money.
Blockchain 2.0: Smart Contracts
The main issues that came with Bitcoin are wasteful mining and lack of network scalability.
To overcome these issues, this version extends the concept of Bitcoin beyond currency. The
new key concepts are Smart Contracts. It is small computer programs that "live" in
the blockchain. They are free computer programs which executed automatically and checked
conditions which are defined earlier like facilitation, verification or enforcement. The big
advantage of this technology that blockchain offers, making it impossible to tamper or hack
Smart Contracts.
Quickly, the blockchain 2.0 version is successfully processing a high number of daily
transactions on a public network, where millions were raised through ICO (Initial Coin
Offerings), and the market cap increased rapidly.
Blockchain 3.0: DApps
DApps is also known as a decentralized application. It uses decentralized storage and
communication. Its backend code is running on a decentralized peer-to-peer network. A
DApp can have frontend code hosted on decentralized storages such as Ethereum Swarm
and user interfaces written in any language that can make a call to its backend like a
traditional Apps.
Blockchain Key Areas
In the blockchain technology, bitcoin is the best-known implementation of the blockchain.
There is a lot of development and the direction is based on the premise of what blockchain
does to enable Bitcoin to happen. We can learn and expand how it can spread into so many
different areas.
The blockchain technology
fixes three things that the Internet was not
designed to do. They are
1. Value
2. Trust
3. Reliability
Value
With blockchain, you can actually create value on a digital asset. The value can be controlled by
that person who owns it. It enables a unique asset to be transferred over the internet without a
middle centralized agent.
Trust
Blockchain enables to securely assign ownership of a specific digital asset and be able to track
who actually controls that asset at a time. In other words, blockchain creates a permanent, secure,
unalterable record of who owns what. It uses advanced hash cryptography to preserve the
integrity of the information.
Reliability
Blockchain distributes their workload among thousands of different computers worldwide. It
provides reliability because if you have everything localized in one location, it becomes a single
point of failure. But, its decentralized network structure ensures that there is no single point of
failure which could bring the entire system down.
Limitation of Blockchain Technology
Blockchain technology has enormous potential in creating trustless, decentralized applications.
But it is not perfect. There are certain barriers which make the blockchain technology not the
right choice and unusable for mainstream application. We can see the limitations of blockchain
technology in the following image.

Lack of Awareness
There is a lot of discussion about blockchain, but people do not know the true value of
blockchain and how they could implement it in different situations.
Limited availability of technical talent
Today, there are a lot of developers available who can do a lot of different things in every field.
But in the blockchain technology, there are not so many developers available who have
specialized expertise in blockchain technology. Hence, the lack of developers is a hindrance to
developing anything on the blockchain.
Immutable
In immutable, we cannot make any modifications to any of the records. It is very helpful if you
want to keep the integrity of a record and make sure that nobody ever tampers with it. But
immutability also has a drawback.
Key Management
As we know, blockchain is built on cryptography, which implies that there are different keys,
such as public keys and private keys. When you are dealing with a private key, then you are also
running the risk that somebody may lose access to your private key. It happens a lot in the early
days when bitcoin wasn't worth that much. People would just collect a lot of bitcoin, and then
suddenly forgot what the key was, and those may be worth millions of dollars today.
Scalability
Blockchain like bitcoin has consensus mechanisms which require every participating node to
verify the transaction. It limits the number of transactions a blockchain network can process. So
bitcoin was not developed to do the large scale volumes of transactions that many of the other
institutions are doing. Currently, bitcoin can process a maximum of seven transactions per
second.
Consensus Mechanism
In the blockchain, we know that a block can be created in every 10 minutes. It is because every
transaction made must ensure that every block in the blockchain network must reach a common
consensus. Depending on the network size and the number of blocks or nodes involved in a
blockchain, the back-and-forth communications involved to attain a consensus can consume a
considerable amount of time and resources.

Crypto Currency
What Is Cryptocurrency?
A cryptocurrency is a digital or virtual currency that is secured by cryptography, which makes it
nearly impossible to counterfeit or double-spend. Many cryptocurrencies are decentralized
networks based on blockchain technology—a distributed ledger enforced by a disparate network
of computers. A defining feature of cryptocurrencies is that they are generally not issued by any
central authority, rendering them theoretically immune to government interference or
manipulation.
Cryptocurrency – meaning and definition
Cryptocurrency, sometimes called crypto-currency or crypto, is any form of currency that exists
digitally or virtually and uses cryptography to secure transactions. Cryptocurrencies don't have a
central issuing or regulating authority, instead using a decentralized system to record transactions
and issue new units.
What is cryptocurrency?
Cryptocurrency is a digital payment system that doesn't rely on banks to verify transactions. It’s
a peer-to-peer system that can enable anyone anywhere to send and receive payments. Instead of
being physical money carried around and exchanged in the real world, cryptocurrency payments
exist purely as digital entries to an online database describing specific transactions. When you
transfer cryptocurrency funds, the transactions are recorded in a public ledger. Cryptocurrency is
stored in digital wallets.
Cryptocurrency received its name because it uses encryption to verify transactions. This means
advanced coding is involved in storing and transmitting cryptocurrency data between wallets and
to public ledgers. The aim of encryption is to provide security and safety.
The first cryptocurrency was Bitcoin, which was founded in 2009 and remains the best known
today. Much of the interest in cryptocurrencies is to trade for profit, with speculators at times
driving prices skyward.
How does cryptocurrency work?
Cryptocurrencies run on a distributed public ledger called blockchain, a record of all transactions
updated and held by currency holders.
Units of cryptocurrency are created through a process called mining, which involves using
computer power to solve complicated mathematical problems that generate coins.
Users can also buy the currencies from brokers, then store and spend them using cryptographic
wallets. If you own cryptocurrency, you don’t own anything tangible. What you own is a key
that allows you to move a record or a unit of measure from one person to another without a
trusted third party.
Blockchain
Central to the appeal and functionality of Bitcoin and other cryptocurrencies is blockchain
technology. As its name indicates, blockchain is essentially a set of connected blocks or an
online ledger. Each block contains a set of transactions that have been independently verified
by each member of the network. Every new block generated must be verified by each node
before being confirmed, making it almost impossible to forge transaction histories.
Types of Cryptocurrency
Bitcoin is the most popular and valuable cryptocurrency. An anonymous person called Satoshi
Nakamoto invented it and introduced it to the world via a white paper in 2008. There are
thousands of cryptocurrencies present in the market today.
Non-Bitcoin cryptocurrencies are collectively known as “altcoins” to distinguish them from the
original.
Each cryptocurrency claims to have a different function and specification. For
example, Ethereum's ether markets itself as gas for the underlying smart
contract platform. Ripple's XRP is used by banks to facilitate transfers between different
geographies.
In the wake of Bitcoin's success, many other cryptocurrencies, known as "altcoins," have been
launched. Some of these are clones or forks of Bitcoin, while others are new currencies that
were built from scratch. They include Solana, Litecoin, Ethereum, Cardano, and EOS. By
November 2021, the aggregate value of all the cryptocurrencies in existence had reached over
$2.1 trillion—Bitcoin represented approximately 41% of that total value.
Exampls of Cryptocurrencies
There are thousands of cryptocurrencies. Some of the best known include:
Bitcoin:
Founded in 2009, Bitcoin was the first cryptocurrency and is still the most commonly traded.
The currency was developed by Satoshi Nakamoto – widely believed to be a pseudonym for an
individual or group of people whose precise identity remains unknown.
Ethereum:
Developed in 2015, Ethereum is a blockchain platform with its own cryptocurrency, called Ether
(ETH) or Ethereum. It is the most popular cryptocurrency after Bitcoin.
Litecoin:
This currency is most similar to bitcoin but has moved more quickly to develop new innovations,
including faster payments and processes to allow more transactions.
Ripple:
Ripple is a distributed ledger system that was founded in 2012. Ripple can be used to track
different kinds of transactions, not just cryptocurrency. The company behind it has worked with
various banks and financial institutions.
How to buy cryptocurrency
There are typically three steps involved in buying cryptocurrency safely. These are:
Step 1: Choosing a platform
The first step is deciding which platform to use. Generally, you can choose between a traditional
broker or dedicated cryptocurrency exchange:
 Traditional brokers. These are online brokers who offer ways to buy and sell
cryptocurrency, as well as other financial assets like stocks, bonds, and ETFs. These
platforms tend to offer lower trading costs but fewer crypto features.
 Cryptocurrency exchanges. There are many cryptocurrency exchanges to choose from,
each offering different cryptocurrencies, wallet storage, interest-bearing account options,
and more. Many exchanges charge asset-based fees.
When comparing different platforms, consider which cryptocurrencies are on offer, what fees
they charge, their security features, storage and withdrawal options, and any educational
resources.
Step 2: Funding your account
Once you have chosen your platform, the next step is to fund your account so you can begin
trading. Most crypto exchanges allow users to purchase crypto using fiat (i.e., government-
issued) currencies such as the US Dollar, the British Pound, or the Euro using their debit or
credit cards – although this varies by platform.
Crypto purchases with credit cards are considered risky, and some exchanges don't support them.
Some credit card companies don't allow crypto transactions either. This is because
cryptocurrencies are highly volatile, and it is not advisable to risk going into debt — or
potentially paying high credit card transaction fees — for certain assets.
Some platforms will also accept ACH transfers and wire transfers. The accepted payment
methods and time taken for deposits or withdrawals differ per platform. Equally, the time taken
for deposits to clear varies by payment method.
An important factor to consider is fees. These include potential deposit and withdrawal
transaction fees plus trading fees. Fees will vary by payment method and platform, which is
something to research at the outset.
Step 3: Placing an order
You can place an order via your broker's or exchange's web or mobile platform. If you are
planning to buy cryptocurrencies, you can do so by selecting "buy," choosing the order type,
entering the amount of cryptocurrencies you want to purchase, and confirming the order. The
same process applies to "sell" orders.
There are also other ways to invest in crypto. These include payment services like PayPal,
Cash App, and Venmo, which allow users to buy, sell, or hold cryptocurrencies. In addition,
there are the following investment vehicles:
 Bitcoin trusts: You can buy shares of Bitcoin trusts with a regular brokerage account.
These vehicles give retail investors exposure to crypto through the stock market.
 Bitcoin mutual funds: There are Bitcoin ETFs and Bitcoin mutual funds to choose
from.
 Blockchain stocks or ETFs: You can also indirectly invest in crypto through blockchain
companies that specialize in the technology behind crypto and crypto transactions.
Alternatively, you can buy stocks or ETFs of companies that use blockchain technology.
The best option for you will depend on your investment goals and risk appetite.
How to store cryptocurrency
Once you have purchased cryptocurrency, you need to store it safely to protect it from hacks or
theft. Usually, cryptocurrency is stored in crypto wallets, which are physical devices or online
software used to store the private keys to your cryptocurrencies securely. Some exchanges
provide wallet services, making it easy for you to store directly through the platform. However,
not all exchanges or brokers automatically provide wallet services for you.
There are different wallet providers to choose from. The terms “hot wallet” and “cold wallet” are
used:
 Hot wallet storage: "hot wallets" refer to crypto storage that uses online software to
protect the private keys to your assets.
 Cold wallet storage: Unlike hot wallets, cold wallets (also known as hardware wallets)
rely on offline electronic devices to securely store your private keys.
Typically, cold wallets tend to charge fees, while hot wallets don't.
What can you buy with cryptocurrency?
When it was first launched, Bitcoin was intended to be a medium for daily transactions, making
it possible to buy everything from a cup of coffee to a computer or even big-ticket items like real
estate. That hasn’t quite materialized and, while the number of institutions accepting
cryptocurrencies is growing, large transactions involving it are rare. Even so, it is possible to buy
a wide variety of products from e-commerce websites using crypto. Examples:
Technology and e-commerce sites:
Several companies that sell tech products accept crypto on their websites, such as newegg.com,
AT&T, and Microsoft. Overstock, an e-commerce platform, was among the first sites to accept
Bitcoin. Shopify, Rakuten, and Home Depot also accept it.
Luxury goods:
Some luxury retailers accept crypto as a form of payment. For example, online luxury retailer
Bitdials offers Rolex, Patek Philippe, and other high-end watches in return for Bitcoin.
Cars:
Some car dealers – from mass-market brands to high-end luxury dealers – already accept
cryptocurrency as payment.
Insurance:
In April 2021, Swiss insurer AXA announced that it had begun accepting Bitcoin as a mode of
payment for all its lines of insurance except life insurance (due to regulatory issues). Premier
Shield Insurance, which sells home and auto insurance policies in the US, also accepts Bitcoin
for premium payments.
If you want to spend cryptocurrency at a retailer that doesn’t accept it directly, you can use a
cryptocurrency debit card, such as BitPay in the US.
Cryptocurrency fraud and cryptocurrency scams
Unfortunately, cryptocurrency crime is on the rise. Cryptocurrency scams include:
Fake websites: Bogus sites which feature fake testimonials and crypto jargon promising
massive, guaranteed returns, provided you keep investing.
Virtual Ponzi schemes: Cryptocurrency criminals promote non-existent opportunities to invest
in digital currencies and create the illusion of huge returns by paying off old investors with new
investors’ money. One scam operation, BitClub Network, raised more than $700 million before
its perpetrators were indicted in December 2019.
"Celebrity" endorsements: Scammers pose online as billionaires or well-known names who
promise to multiply your investment in a virtual currency but instead steal what you send. They
may also use messaging apps or chat rooms to start rumours that a famous businessperson is
backing a specific cryptocurrency. Once they have encouraged investors to buy and driven up the
price, the scammers sell their stake, and the currency reduces in value.
Romance scams: The FBI warns of a trend in online dating scams, where tricksters persuade
people they meet on dating apps or social media to invest or trade in virtual currencies. The
FBI’s Internet Crime Complaint Centre fielded more than 1,800 reports of crypto-focused
romance scams in the first seven months of 2021, with losses reaching $133 million.
Otherwise, fraudsters may pose as legitimate virtual currency traders or set up bogus exchanges
to trick people into giving them money. Another crypto scam involves fraudulent sales pitches
for individual retirement accounts in cryptocurrencies. Then there is straightforward
cryptocurrency hacking, where criminals break into the digital wallets where people store their
virtual currency to steal it.
Four tips to invest in cryptocurrency safely
According to Consumer Reports, all investments carry risk, but some experts consider
cryptocurrency to be one of the riskier investment choices out there. If you are planning to invest
in cryptocurrencies, these tips can help you make educated choices.
Research exchanges:
Before you invest, learn about cryptocurrency exchanges. It’s estimated that there are over 500
exchanges to choose from. Do your research, read reviews, and talk with more experienced
investors before moving forward.
Know how to store your digital currency:
If you buy cryptocurrency, you have to store it. You can keep it on an exchange or in a digital
wallet. While there are different kinds of wallets, each has its benefits, technical requirements,
and security. As with exchanges, you should investigate your storage choices before investing.
Diversify your investments:
Diversification is key to any good investment strategy, and this holds true when you are
investing in cryptocurrency. Don't put all your money in Bitcoin, for example, just because that's
the name you know. There are thousands of options, and it's better to spread your investment
across several currencies.
Prepare for volatility:
The cryptocurrency market is highly volatile, so be prepared for ups and downs. You will see
dramatic swings in prices. If your investment portfolio or mental wellbeing can't handle that,
cryptocurrency might not be a wise choice for you.
Cryptocurrency is all the rage right now, but remember, it is still in its relative infancy and is
considered highly speculative. Investing in something new comes with challenges, so be
prepared. If you plan to participate, do your research, and invest conservatively to start.
One of the best ways you can stay safe online is by using a comprehensive antivirus. Kaspersky
Internet Security defends you from malware infections, spyware, data theft and protects your
online payments using bank-grade encryption.
Advantages and Disadvantages of Cryptocurrency
Cryptocurrencies were introduced with the intent to revolutionize financial infrastructure. As
with every revolution, however, there are tradeoffs involved. At the current stage of
development for cryptocurrencies, there are many differences between the theoretical ideal of a
decentralized system with cryptocurrencies and its practical implementation.
Advantages
 Cryptocurrencies represent a new, decentralized paradigm for money. In this system,
centralized intermediaries, such as banks and monetary institutions, are not necessary to
enforce trust and police transactions between two parties. Thus, a system with
cryptocurrencies eliminates the possibility of a single point of failure.
 Cryptocurrencies promise to make it easier to transfer funds directly between two
parties, without the need for a trusted third party like a bank or a credit card company.
Such decentralized transfers are secured by the use of public keys and private keys and
different forms of incentive systems, such as proof of work or proof of stake.
 Because they do not use third-party intermediaries, cryptocurrency transfers between
two transacting parties are faster as compared to standard money transfers.
 Cryptocurrency investments can generate profits. Cryptocurrency markets have
skyrocketed in value over the past decade, at one point reaching almost $2 trillion.
 Bitcoin serve as intermediate currencies to streamline money transfers across borders.
Thus, a fiat currency is converted to Bitcoin, transferred across borders and,
subsequently, converted to the destination fiat currency. This method streamlines the
money transfer process and makes it cheaper.
Disadvantages
 Cryptocurrencies are actually pseudonymous. They leave a digital trail that agencies
such as the Federal Bureau of Investigation (FBI) can decipher. This opens up
possibilities of governments or federal authorities tracking the financial transactions of
ordinary citizens.
 Cryptocurrencies have become a popular tool with criminals for nefarious activities such
as money laundering and illicit purchases. Cryptocurrencies have also become a favorite
of hackers who use them for ransomware activities.
 In theory, cryptocurrencies are meant to be decentralized, their wealth distributed
between many parties on a blockchain. In reality, ownership is highly concentrated.
 One of the conceits of cryptocurrencies is that anyone can mine them using a computer
with an Internet connection. However, mining popular cryptocurrencies requires
considerable energy, sometimes as much energy as entire countries consume.
 Though cryptocurrency blockchains are highly secure, other crypto repositories, such as
exchanges and wallets, can be hacked.
 Cryptocurrencies traded in public markets suffer from price volatility. Bitcoin has
experienced rapid surges and crashes in its value.
Quantum Computing
It is the use of quantum mechanics to run calculations on specialised hardware.
Quantum computing is an area of computing focused on developing computer technology based
on the principles of quantum theory (which explains the behavior of energy and material on the
atomic and subatomic levels).
Computers used today can only encode information in bits that take the value of 1 or 0—
restricting their ability. On the other hand, Quantum computing uses quantum bits or qubits. It
harnesses the unique ability of subatomic particles that allows them to exist in more than one
state (i.e., a 1 and a 0 at the same time).
To fully define quantum computing, we need to define some key terms first.
What is quantum?
The quantum in "quantum computing" refers to the quantum mechanics that the system uses to
calculate outputs. In physics, a quantum is the smallest possible discrete unit of any physical
property. It usually refers to properties of atomic or subatomic particles, such as electrons,
neutrinos and photons.
What is a qubit?
A qubit is the basic unit of information in quantum computing. Qubits play a similar role in
quantum computing as bits play in classical computing, but they behave very differently.
Classical bits are binary and can hold only a position of 0 or 1, but qubits can hold a
superposition of all possible states.
What Is Quantum Computing?
Quantum computers harness the unique behaviour of quantum physics—such as
 Superposition,
 Entanglement and
 Quantum interference
and apply it to computing.
 Superposition
In superposition, quantum particles are a combination of all possible states. They fluctuate until
they are observed and measured. One way to picture the difference between binary position and
superposition is to imagine a coin. Classical bits are measured by "flipping the coin" and getting
heads or tails. However, if you were able to look at a coin and see both heads and tails at the
same time, as well as every state in between, the coin would be in superposition.

 Entanglement
Entanglement is the ability of quantum particles to correlate their measurement results with each
other. When qubits are entangled, they form a single system and influence each other. We can
use the measurements from one qubit to draw conclusions about the others. By adding and
entangling more qubits in a system, quantum computers can calculate exponentially more
information and solve more complicated problems.
 Quantum interference
Quantum interference is the intrinsic behaviour of a qubit, due to superposition, to influence the
probability of it collapsing one way or another. Quantum computers are designed and built to
reduce interference as much as possible and ensure the most accurate results. To this end,
Microsoft uses topological qubits, which are stabilised by manipulating their structure and
surrounding them with chemical compounds that protect them from outside interference.

The Uses
Quantum applications are transforming how we live, work and play. Technologies like
quantum sensors, quantum computers and quantum information security are emerging
from labs around the world, and we are already seeing the tremendous possibilities.

Quantum computing could contribute greatly in the fields of finance, military affairs and
intelligence, drug design and discovery, aerospace designing, utilities (nuclear fusion), polymer
design, machine learning and artificial intelligence (AI) and Big Data search, and digital
manufacturing.
How does quantum computing work?
A quantum computer has three primary parts:
 An area that houses the qubits
 A method for transferring signals to the qubits
 A classical computer to run a program and send instructions
For some methods of qubit storage, the unit that houses the qubits is kept at a temperature just
above absolute zero to maximise their coherence and reduce interference. Other types of qubit
housing use a vacuum chamber to help minimise vibrations and stabilise the qubits.
Signals can be sent to the qubits using a variety of methods, including microwaves, laser and
voltage.
Quantum computer uses and application areas
A quantum computer cannot do everything faster than a classical computer, but there are a few
areas where quantum computers have the potential to make a big impact.
Quantum simulation
Quantum computers work exceptionally well for modelling other quantum systems because they
use quantum phenomena in their computation. This means that they can handle the complexity
and ambiguity of systems that would overload classical computers. Examples of quantum
systems that we can model include photosynthesis, superconductivity and complex molecular
formations.
Cryptography
Classical cryptography—such as the Rivest–Shamir–Adleman (RSA) algorithm that is widely
used to secure data transmission—relies on the intractability of problems such as integer
factorisation or discrete logarithms. Many of these problems can be solved more efficiently using
quantum computers.
Optimisation
Optimisation is the process of finding the best solution to a problem given its desired outcome
and constraints. In science and industry, critical decisions are made based on factors such as cost,
quality and production time—all of which can be optimised. By running quantum-inspired
optimisation algorithms on classical computers, we can find solutions that were previously
impossible. This helps us find better ways to manage complex systems such as traffic flows,
airplane gate assignments, package deliveries and energy storage.
Quantum machine learning
Machine learning on classical computers is revolutionising the world of science and business.
However, training machine learning models comes with a high computational cost and that has
hindered the scope and development of the field. To speed up progress in this area, we are
exploring ways to devise and implement quantum software that enables faster machine learning.
Search
A quantum algorithm developed in 1996 dramatically sped up the solution to unstructured data
searches, running the search in fewer steps than any classical algorithm could.
Real-World Example of a Quantum Computer
Google (GOOG) is spending billions of dollars on its plan to build its quantum computer by
2029. The company has opened a campus in California, called Google AI, to help it meet its
goal. Google has been investing in this technology for years. As well, so have other companies,
such as Honeywell International (HON) and International Business Machine (IBM). IBM
expects to hit major quantum computing milestones in the coming years.

You might also like