Information System Ioe Notes
Information System Ioe Notes
• Data is the raw fact that has no any significance and use if not processed.
• Data is processed to generate information.
• Information is those processed data that provides some meaning and can be
used for decision making and organizational improvements.
Types of information
1. Operational Information:
• Those information that are needed to perform the daily operations of an
organization are called operational information.
• Eg: daily account of sales and stocks, low stock item list, profit and loss
account and so on.
• These information are used by line managers to facilitate proper survival of the
organization by taking appropriate decisions that would facilitates the user’s
expectations and needs.
2. Tactical Information:
• Those information that are used to generate tactics to facilitate improvement in
profit and performance of an organization.
• It is generally collected on weekly or monthly basis.
• Eg: Whether to provide discount on some stocks.
• It is used by middle managers.
3. Strategic Information:
• Those information needed for long term improvement and planning are called
strategic information.
• It includes information on organizational expansion and new opportunities.
• Such information is generally unstructured.
• It is taken on quarterly or yearly basis.
• It is used by the top level managers.
• It helps in decisions like whether to open a branch, whether to initiate customer
cards, etc.
4. . Statutory Information:
• Those information that are needed to provide to the government about the
organizational structure and status are called statutory information.
• Eg: VAT filling, audits and so on.
• It examines the information against the criteria under the identified purpose of
the audit to meet the standard compliance.
• It determines the user information needs.
• It lists the information resources available within an organization.
• It identifies the costs and benefits of the information resources available.
• It provides information about the working structure of the information system
of an organization.
• It produces report that recommends for the information handling problems.
• It helps organization to make use of information for strategic planning and
implementations.
• It aids in decision making and support.
• It enables organization to be dynamic i.e. adapt to necessary changes.
• Information audit helps to identify problems like data redundancy, duplication,
inconsistency and cost to store and utilize data and information.
• Information audit helps to identify hidden assets of an organization, skills and
expertise of staffs, market for further expansion and so on that would expand
organizational opportunities.
Methodological Approach to Information Audit
• Support of senior management is very crucial for information audit and in most
cases such support is not provided.
• It is difficult to decide whether to use internal auditors or external consultants.
• It is very tedious task to collect and gather necessary information for auditing.
• The information audit time span depends up on the size of an organization.
• It is difficult to establish costs and value of information
Layered Security
SSL
Extended Validation:
• Extended validation is a certificate used for HTTPS websites and software that
proves the legal entity controlling the websites or software package.
• To obtain EV certificate, verification of the requesting entity’s identity is
required by a certificate authority.
• It increases the security due to the identity validation process, which is
indicated within the certificate by the policy identifier.
Telnet:
• Telnet is the TCP/IP protocol standard that allows users to log on remotely and
access resources as if the user had a local terminal connection to the server.
• The major threat of using Telnet is that it uses TCP/IP connection for
information flow that has less security.
• The Telnet uses TCP port 23 for connection.
Secure Shell (SSH)
Components of EMS
Features of ERP:
Benefits of ERP:
1. It integrates all the data in a single source that makes the data and information
to flow easily and transparently across the intended departments.
2. It is a real time system that considers customers as well as suppliers along with
the internal processes as a part of the enterprise that helps to improve internal
as well as external communication and understand about the demands on
timely manner.
3. It helps in increasing productivity of an organization based on the customer
current demands.
4. As all the information is integrated to a single computer system, it minimizes
the operating costs of the information system.
5. It helps to trace the past and present business activities and lay foundation to
improve those activities in future for achieving enterprise progress.
6. It helps the enterprise to survive in this competitive world by gaining trust
from the customers as the customers are directly involved as a part of an
enterprise.
7. It helps to make better use of available resources.
Features of SCM:
Challenges of SCM:
1. SCM is not able to perform all the functions that an organization hopes for.
2. Customer demands for the product immediately which is difficult to handle by
the SCM.
3. The delivery of the product to the end customer is difficult for which an
organization should collaborate with others.
4. If SCM system is not reliable, an organization may face problems to track
about their stocks and further productions.
Features of CRM
1. 1. Marketing Automation:
• It is responsible to enhance marketing efforts to customers.
• For eg: the CRM system can automatically send the customers with the
marketing materials via social media or email.
2. Sales Force Automation:
• It is responsible to track all the contacts and follow ups between customer and
sales person so as to prevent duplicate efforts.
3. Contact Center Automation:
• It is responsible to automate the customer contact with the company.
• It can be implemented via a recorded audio or through chat bots.
4. Location Based Services:
• The CRM system include technology that can create geographic marketing
campaigns based on customer's physical locations.
Benefits of CRM
Methodologies
The formal methodologies and methods used to offer organization reusable business
process solutions are as follows:
1. Computer Integrated Manufacturing Open Systems Architecture (It provides
templates to encode business, people and information technology of an
enterprise requirements.)
2. Integrated Definition (It shows business process flows through a variety of
decomposed business functions with corresponding information inputs, outputs
and actors.)
3. Petri Nets (It is used to model manufacturing systems and provide formalisms
for the modeling of concurrent systems with the ability to create simple state
representation, concurrent system transitions and allow duration of transitions.)
4. Unified Enterprise Modeling Language (It is the object oriented enterprise
modeling tool in which emphasis is placed on the usage of enterprise objects
from which complex enterprise systems are made.)
5. Enterprise Function Diagrams (It is a modeling technique for the representation
of enterprise functions and their corresponding interactions. It provides easy to
use and detailed representation about a business process and its corresponding
functions, inputs, outputs and triggers.)
Loose Integration
Tight Integration
• Tight integration is the system that works on the inter connectivity and inter-
processing of two or more systems simultaneously to deliver a
cohesive/integrated solution.
• The entire logic is distributed across several hardware and software
components, which all need to be operational and connected to deliver the
business logic/process.
• For example, a bank ATM machine depends on the ATM machine hardware,
built-in firmware/applications and the primary banking application to allow a
customer to withdraw cash or access any ATM-specific services.
1. Passive DSS
• It is a system that aids the process of decision making but that can not bring the
explicit decision solutions.
2. Active DSS
• It is a system that is able to bring explicit decision solutions for a problem.
3. Cooperative DSS
• It is a system that allows iterative process between human and system to
achieve best solution.
Components of DSS
1. Database:
• It is a well organized collection of current and historical data from a number of
applications and groups.
• It provides easy access to data and information from variety of applications.
• It is able to maintain data integrity.
• The data are extracted from various relevant databases and stored especially
for decision support system only.
2. Model:
• A model represents an abstract representation of different components and
relationships of a phenomenon.
• The model can be classified as:
(a) Behavioral Model
(b) Management Science Model
(c) Operation Research Model
3. User Interface:
• It allows interaction between the user of the system and the DSS database and
model.
• The user interface is responsible to deal with the end user of the system
providing them with friendly interface.
Behavioral Model:
DSS MIS
• DSS provides support for • MIS provides support for
unstructured or semi structured structured decision making.
decision making.
• MIS makes use of quantitative
• DSS makes use of qualitative data for data for decision making.
decision making.
• MIS focuses on information
• DSS focuses on leadership and gathered and report planning.
decision making.
• MIS helps to achieve operational
• DSS helps an organization to choose efficiency.
right path for its progress.
• Flow of information is on both
• Flow of information is only upward sides (up and down) in case of
in case of DSS. MIS.
4.2 Group Decision Support System (GDSS)
• Group Decision Support System is an interactive computer based system that
facilitates a number of decision-makers (working together in a group) in
finding solutions to problems that are unstructured in nature.
• GDSS takes inputs from multiple users interacting simultaneously with the
systems to arrive at a decision as a group.
• It improves the quality and effectiveness of the group meetings.
• It reduces time of high level managers to reach at efficient decisions.
Components of GDSS
1. Hardware:
• It includes electronic hardware like computer, equipment used for networking,
electronic display boards and audio visual equipment.
• It also consists of conference infrastructures.
• All these hardware facilitates the support for group decision support system.
2. Software Tools:
• It includes various tools and techniques that helps the decision makers to plan,
organize ideas, gather information, establish priorities, take decisions and
document the meeting proceedings.
• The tools and techniques consist of electronic questionnaire, idea organizer,
priority setting tools, policy formation tools, modeling tools, dialogue manager,
database management system and so on.
3. People:
• It includes members to participate in the meeting, a trained facilitator who
helps with the proceedings of the meeting, and an expert staff to support the
hardware and software.
Features of GDSS
1. It is easy and simple to use because of its interactive and user friendly user
interface.
2. It facilitates users at different locations to make decisions as a group that
results in better decisions.
3. The facilitator provides general support to the group and helps them to use the
system.
4. It emphasizes on unstructured and semi structured decisions.
5. It supports all phases of decision making (intelligence, design, choice and
implementation).
6. The participants are able to share their ideas more openly.
Characteristics of ESS:
1. Informational characteristics
i. Flexibility and ease of use.
ii. Provides the timely information with the short response time and also with
the quick retrieval.
iii. Produces the correct information.
iv. Produces the relevant information.
v. Produces the validated information.
2. User interface/orientation characteristics
i. Consists of the sophisticated self help.
ii. Contains the user friendly interfaces consisting of the graphic user.
iii. Can be used from many places.
iv. Offers secure reliable, confidential access along with the access procedure.
v. Is very much customized.
vi. Suites the management style of the individual executives.
3. Managerial / executive characteristics
i. Supports the over all vision, mission and the strategy.
ii. Provides the support for the strategic management.
iii. Sometimes helps to deal with the situations that have a high degree of risk.
iv. Is linked to the value added business processes.
v. Supports the need/ access for/ to the external data/ databases.
vi. Is very much result oriented in the nature.
Benefits of ESS:
Features
a) Reasoning capacity
b) Cope with uncertainty
c) Use of knowledge not data
d) Symbolic knowledge representation
e) Use meta knowledge
f) Use user interface
Virtual Reality
• Data mining is the process of finding patterns in the given set of data.
• It is also defined as the process of extracting information from huge sets of
data.
• Such patterns generally provides some meaningful information to the intended
users.
• It uses statistical analysis.
• The applications of data mining are as follows:
1. Market analysis
2. Fraud detection
3. Customer retention
4. Production control
• Data warehousing is the process of aggregating data from multiple sources into
a common repository.
• Data warehouse is a database which is kept separate from the operational
database, which basically consists of consolidated historical data.
• A data warehouse helps executives to organize, understand, and use their data
to take strategic decisions.
• The features of a data warehouse are as follows:
1. Subject oriented
2. Integrated
3. Time variant
4. Non-volatile
• Data warehousing is the process in which the data from different sources are
aggregated to a single database.
• Data mining is the process of analyzing the data stored in the data warehouse
to generate some meaningful patterns.
• Data mining is the process that is incomplete without data warehousing.
• The general flow of data analysis is:
Data from different source ====> Data warehouse =====> Data mining
=====> Useful patterns
Necessity of Data Warehousing
A data warehouses is kept separate from operational databases due to the following
reasons:
1. An operational database is constructed for well-known tasks and workloads
such as searching particular records, indexing, etc. In contract, data warehouse
queries are often complex and they present a general form of data.
2. Operational databases support concurrent processing of multiple transactions.
Concurrency control and recovery mechanisms are required for operational
databases to ensure robustness and consistency of the database.
3. An operational database query allows to read and modify operations, while a
data warehouse query needs only read only access of stored data.
4. An operational database maintains current data. On the other hand, a data
warehouse maintains historical data.
Differences:
Inserts and Short and fast inserts and Periodic long-running batch jobs
Updates updates initiated by end users refresh the data
Necessity of IS Planning
Change Management
• It consists of all the nodes that can reach the giant SCC but can not be reached
from it.
• It is the nodes that are upstream of giant SCC.
OUT Component:
• It consists of all the nodes that can be reached from the giant SCC but can not
reach giant SCC.
• It is the nodes that are downstream of giant SCC.
Tendrils:
• It consists of:
1. Nodes reachable from IN component that can not reach the giant SCC.
2. Nodes that can reach OUT but can not be reached from giant SCC.
Disconnected Component:
• It consists of the nodes that would not have any path to the giant SCC even if
we completely ignored the direction of the edges.
• Page rank is the composite score given by the search engines to the web pages
to find and index them when user searches for the query.
• It is the algorithm that is used by the search engines to rank the websites in
their results.
• It works by counting the number and quality of links to a page to determine a
rough estimate of how important the website is.
• Page rank is a link analysis algorithm and it assigns a numerical weighting to
each element of a hyperlinked set of documents, such as the World Wide Web,
with the purpose of "measuring" its relative importance within the set.
• Page rank is computed as follows:
1. In a network with n nodes, we assign all nodes the same initial Page Rank, set
to be
1/n.
2. We choose a number of steps k.
3. We then perform a sequence of k updates to the Page Rank values, using the
following rule for each update:
a) Basic Page Rank Update Rule: Each page divides its current Page Rank
equally
across its out-going links, and passes these equal shares to the pages it points
to. (If a page has no out-going links, it passes all its current Page Rank to
itself.) Each page updates its new Page Rank to be the sum of the shares it
receives.
Search Engine:
• Search engine is the huge database of internet resources that helps to locate
information on the World Wide Web.
• Users can search for any information in a search engine by passing query in the
form of keywords or phrase.
• The query is then searches in its database and the results are displayed back to
the users.
1. The user enters the keyword to search for the required information through a
query in the search engine user interface.
2. The search engine looks for the keyword in the index for predefined database
instead of going directly to the web to search for the keyword.
3. It then uses software to search for the information in the database. This
software component is known as web crawler.
4. Once web crawler finds the pages, the search engine then shows the relevant
web pages as a result. These retrieved web pages generally include title of
page, size of text portion, first several sentences etc.
5. User can click on any of the search results to open it to get the relevant
information.
Web Mining:
• Web mining is the data mining technique that is used to discover patterns from
the World Wide Web.
• It is the process of gathering information by mining (extracting something
useful) the web.
• It is divided into three types:
1. Web Content Mining
2. Web Usage Mining
3. Web Structure Mining
• Web usage mining is the process of extracting patterns and information from
server logs to gain insight on user activity including where the users are from,
how many clicked what item on the site and the types of activities being done
on the site.
• It provides basic insights on how the users are using the web.
• It helps to discover the web usage patterns from the web data to understand and
serve the needs of web based applications.
• Web structure mining is the process of using graph theory to analyze the node
and connection structure of a web site.
• It helps to extract the patterns from the hyperlinks in the web.
• It helps to analyze the document structure to describe the structure of the web
site.
• Web structure mining can be used for page ranking of the web sites for search
engines.
Collaborative Filtering:
1. 1. Cold Start:
• The system requires a huge amount of existing data on a user so as to make
accurate recommendations.
• This problem is termed as cold start.
2. Scalability:
• In the real world system, there are millions of users and products.
• So, to calculate recommendations, a large computational power should be
possessed by the system.
3. Sparsity:
• All the users do not rate the items.
• So, even the most popular items may have few ratings.
1. It is effective to recommend same type of items as the user is using. For eg:
recommending news articles based on browsing of news.
1. Pandora Radio (It plays music with similar characteristics to that of a song
provided by the user as the initial seed)
2. Rotten Tomatoes (Movie recommendation system)
Voluminous Data
Association Rule
• Association rule learning is a method for discovering interesting correlations
between variables in large databases.
• It helps to understand closeness of products with each other so as to increase
sales.
• It can be used to monitor system logs to detect intruders and malicious activity.
• It helps to extract information about visitors to websites from web server logs.
Classification Tree Analysis
• Statistical classification is the method of identifying categories that a new
observation belongs to.
• It requires training set of correctly identified historical data.
• It mainly helps in assigning objects to categories and groups automatically.
Machine Learning
• Machine learning is the method of inducing human like sense to real world
inside the machine.
• It provides ability to the computers to learn without being explicitly
programmed.
• It helps in predictions based on known properties learned from sets of training
data.
Regression Analysis
• Regression analysis is the method that involves manipulating some
independent variable to see how it influences a dependent variable.
• It describes how the value of dependent variable changes when the
independent variable is varied.
• It is used to understand customer satisfaction against loyalty.
Sentiment Analysis
• Sentiment analysis is the method that determine the sentiments (view) of
speakers or writers with respect to a topic.
• It is used in improving services by analyzing guest comments and customers
demands.
Volume:
• Volume indicates the quantity of generated and stored data.
• The size of data determines the potential value and insight.
• It also determines whether it is considered to be big data or not.
• The volume of data in the world is increasing exponentially .
Variety:
• Variety indicates the different types and nature of data.
• All the data present in big data analysis may not be of same type.
• Even a single application may be generating variety of data.
• This increases complexity in big data analysis and knowledge extraction.
• The data may be web data, relational data, XML, structured data, streaming
data, graph data and so on.
• For efficient extraction of information or patterns, all these variety of data must
be linked together and analyzed together.
Velocity:
• Velocity indicates the speed at which data is generated and processed to meet
the demands.
• The data obtained is of dynamic nature, so they must be analyzed very fast to
provide efficient and effective knowledge.
Cloud
Cloud Computing
1. One can access applications over the Internet. It reduces the necessity of
installation of software in the system.
2. One can manipulate and configure applications online at any time using any
devices.
3. It provides tools for online development and deployment.
4. It provides platform independent resources.
5. It operates at high efficiency with optimal utilization.
6. It provides load balancing services.
1. Cloud computing is provided by third party, this may cause risk to handover
the sensitive information to cloud service providers.
2. It is difficult to switch from one cloud service provider to another.
3. The services are accessible by any one from the Internet. So, there may be
compromise if necessary security system is not applied.
4. In some cases, data deletion may be insecure or incomplete.
1. Public Cloud
• It allows systems and services to be easily accessible to the general public.
• It is insecure due to openness characteristics.
2. Private Cloud
• It allows systems and services to be accessible within an organization.
• It is secure because of its private nature.
3. Community Cloud
• It allows systems and services to be accessible by a group of organizations.
4. Hybrid Cloud
• It is the mixture of public and private cloud.
• The sensitive activities are hosted using private cloud.
• The general activities are hosted using public cloud.
Hadoop System
• Hadoop is a framework that allows to process and store huge data sets.
• It is a batch oriented data processing system that works by storing and tracking
data across multiple machines and can scale to thousands of servers.
• It is generally used to process huge data sets that are unstructured in nature.
• The data loaded to Hadoop system is split into pieces and spreads across
different servers.
• It keeps track of where the data is.
• The complex queries can be performed with faster performance as all the
processors are working in parallel.
• For executing such distributed queries, it uses MapReduce.
• It can be divided into two parts: processing and storage.
MapReduce
• MapReduce is the programming model that allows user to process huge data
stored in distributed Hadoop system.
• It provides platform to perform distributed and parallel processing on large
data sets in a distributed environment.
• It consists of tow distinct tasks: Map and Reduce.
• Map task reads and processes a block of data to produce key-value pairs as
intermediate output.
• The output of a Mapper is the input to the Reducer.
• Reduce task receives the key-value pair from multiple map jobs and then
aggregates them into a single result set.
• The single result set is the final output of the system.
The working of MapReduce for word count problem is shown in
figure below:
• Transactional data are those data that needs ACID property to be guaranteed.
• Such data in the cloud is not a perfect match because of following reasons:
1. Cloud provides shared-nothing architecture but transactional data can not be
implemented in such architecture.
2. Since the data are replicated over large geographic distances, it is difficult to
maintain the ACID properties.
3. Storing transactional data on an untrusted cloud host arise a lot of risk of data
compromization.
• Analytical data means those data that are queried up on for use in business
planning, decision support and problem solving.
• The scale of such data is larger as it contains all historical data too.
• Such data are well suited to run in cloud environment due to following reasons:
1. It uses shared-nothing architecture.
2. It does not require ACID properties to be guaranteed.
3. It generally do not contain sensitive data. So, there is no risk of data
compromise.