0% found this document useful (0 votes)
35 views9 pages

Hot Data

Hot data is business-critical information that needs to be accessed quickly and is often stored on fast storage like flash memory. Cold storage is designed for inactive data and prioritizes low cost, high capacity and data durability over speed. Warm data provides an interrelational understanding beyond what can be learned from quantitative data alone.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as RTF, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
35 views9 pages

Hot Data

Hot data is business-critical information that needs to be accessed quickly and is often stored on fast storage like flash memory. Cold storage is designed for inactive data and prioritizes low cost, high capacity and data durability over speed. Warm data provides an interrelational understanding beyond what can be learned from quantitative data alone.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as RTF, PDF, TXT or read online on Scribd
You are on page 1/ 9

HOT DATA

Hot data is data that needs to be accessed frequently. It is typically


business-critical information that needs to be accessed quickly and is often
used by a company for quick decision making.Hot data usually resides on
the fastest storage -- typically flash in hybrid or tiered storage
environments.

Case studies

Public opinion

In the public sectors, the major confrontations are the amalgamation and
ability of the big data from corner to corner of various public sector units
and allied unions. Big data provides a large range of facilities to the
government sectors including the power investigation, deceit recognition,
fitness interconnected exploration, economic promotion investigation and
ecological fortification. Big data is even used to examine the food based
infections by the FDA. Big data results are fast which outputs to quicker
well-being. Also in the investigation of a huge volume of communal
complaints uses the big data analytics. This same analytics are utilized in the
course of health check statistics in urgency and resourcefully for quicker
pronouncement manufacture and to become aware of mistrustful or
falsified declarations.

COLD DATA

Cold storage is a computer system or mode of operation designed for the


retention of inactive data.

High-performance primary storage is generally considered too expensive for


inactive data that is retained on a long-term or indefinite basis. Design
priorities for cold storage may include low cost, high capacity and data
durability. Data retrieval and response time can be significantly slower for a
cold storage system than for devices or systems designed for active data.
Common media choices include Linear Tape-Open (LTO) tape and
commodity hard disk drives (HDDs).

Development of cold data

Public cloud services targeting cold storage include Amazon Glacier, Google
Cloud Storage Nearline and Microsoft Azure Cool Blob Storage. These
services offer lower data availability at a lower price per gigabyte than
alternative cloud storage offerings sold by Amazon, Google and Microsoft.

The Open Compute Project (OCP), driven by Facebook, includes initiatives


dedicated to cold storage. The OCP storage committee focuses on the
development of specifications and software intended to drive innovation in
hardware and to facilitate greater efficiency and value. The OCP project to
create high-capacity, low-cost cold storage identified shingled magnetic
recording (SMR) HDDs with spin-down capability as a highly suitable and
cost-effective technology for cold storage. Facebook has also tested Blu-ray
discs and expressed interest in low-endurance flash for cold
storage.Facebook has built cold storage facilities as part of its data centers
in Prineville, Oregon, and Forest City, North Carolina. Design priorities have
included low power consumption, system efficiency, high density, scalability,
data durability and future-proofing. Facebook has discussed its use of
erasure coding for space-efficient data protection.

WARM DATA

“Warm Data” is information about the interrelationships that integrate


elements of a complex system. It has found the qualitative dynamics and
offers another dimension of understanding to what is learned through
quantitative data, (cold data). Warm Data will provide leverage in our
analysis of other streams of information. The implications for the uses of
Warm Data are staggering, and may offer a whole new dimension to the
tools of information science we have to work with at present.

WHY WARM DATA IS IMPORTANT

In order to interface with any a complex system without disrupting the


circuitry of the interdependencies that give it its integrity we must look at
the spread of relationships that make the system robust. Using only analysis
of statistical data will offer conclusions that can point to actions that are out
of sync with the complexity of the situation. Information without
interrelationality is likely to lead us toward actions that are misinformed,
thereby creating further destructive patterns.

WHAT WARM DATA EFFECTS

Everything. Business, medical, economics, development, ecological


understanding, cross cultural diplomacy, and more. The IBI will make this
data available to political, cultural, economic, educational, and arts sectors.

OPEN DATA

Open data is the idea that some data should be freely available to everyone
to use and republish as they wish, without restrictions from copyright,
patents or other mechanisms of control.The goals of the open data
movement are similar to those of other "open" movements such as open
source, open hardware, open content and open access. The philosophy
behind open data has been long established (for example in the Mertonian
tradition of science), but the term "open data" itself is recent, gaining
popularity with the rise of the Internet and World Wide Web.

Open data may include non-textual material such as maps, genomes,


connectomes, chemical compounds, mathematical and scientific formulae,
medical data and practice, bioscience and biodiversity. Problems often arise
because these are commercially valuable or can be aggregated into works
of value. Access to, or re-use of, the data is controlled by organisations,
both public and private. Control may be through access restrictions,
licenses, copyright, patents and charges for access or re-use. Advocates of
open data argue that these restrictions are against the communal good and
that these data should be made available without restriction or fee. In
addition, it is important that the data are re-usable without requiring
further permission, though the types of re-use.
Example

In governments

Tunisian Open Data Portal for Energy & Mines / data.industrie.gov.tn,


launched in June 2015

Ghana Open Data Initiative, launched in January 2012

Japan Open Data Initiative, launched in December 2013

Jamaica Open Data Portal launched on Friday, June 24, 2016

Kenya Open Data Portal, launched in July 2011

United Kingdom Data.gov.uk platform, launched in January 2010

United States Open Government Initiative, which operates at the federal


level.

University of Babylon Open Access Repository (Arabic), and


Academic/Personnel Data Collections

Singapore government's one-stop portal to publicly-available datasets from


public agencies, launched in 2011.

In science

Examples of open data in science:

The Dataverse Network Project – archival repository software promoting


data sharing, persistent data citation, and reproducible research

data.uni-muenster.de – Open data about scientific artifacts from University


of Muenster, Germany. Launched in 2011.

linkedscience.org/data – Open scientific datasets encoded as Linked Data.


Launched in 2011.

CLOSED DATA
Several mechanisms restrict access to or reuse of data. They include:

·0 making data available for a charge.

·1 compilation in databases or websites to which only registered


members or customers can have access.

·2 use of a proprietary or closed technology or encryption which creates


a barrier for access.

·3 copyright forbidding (or obfuscating) re-use of the data, including the


use of "no derivatives" requirements.

·4 patent forbidding re-use of the data (for example the 3-dimensional


coordinates of some experimental protein structures have been
patented).

·5 restriction of robots to websites, with preference to certain search


engines.

·6 aggregating factual data into "databases" which may be covered by


"database rights" or "database directives" (e.g. Directive on the legal
protection of databases).

·7 time-limited access to resources such as e-journals (which on


traditional print were available to the purchaser indefinitely).

·8 "webstacles", or the provision of single data points as opposed to


tabular queries or bulk downloads of data sets.

·9 political, commercial or legal pressure on the activity of organisations


providing Open Data (for example the American Chemical Society
lobbied the US Congress to limit funding to the National Institutes of
Health for its Open PubChem data).

Example

protools,logic pro,cubase,etc..
THICK DATA

Thick data is qualitative information that provides insights into the everyday
emotional lives of consumers. It goes beyond big data to explain why
consumers have certain preferences, the reasons they behave the way they
do, why certain trends stick and so on.

It’s crucial for successful companies to analyze the emotional way in which
people use their products or services to develop a better understanding of
their customers. By using thick data, companies can develop a positive
relationship with their customers and it becomes easier for those
companies to maintain happy customers and attract new ones.

Example

Big data will tell you that in 2013, Samsung was able to sell 35 million more
smartphones than Apple. But what can these companies really do with this
data? Pat themselves on the back or hang their heads in shame? If you are
in the market for a smartphone, you’re not going to buy a Samsung because
they sold 35 million more than Apple. As a customer, you probably don’t
even know this information. You may, however, buy a Samsung because
they offer a multitude of models that you can customize to your
preferences, and Apple’s product line is less diverse. Or perhaps you won’t
buy an Apple smartphone because it’s not quite as durable, or they don’t
have as wide a selection of phone colors as Samsung. Using thick data to
figure out why more people are buying from Samsung is key for both
companies to move forward and either keep dominating the market, or
reinvent to gain dominance. At its core, business is about making bets on
human behavior, and those bets backed by thick data are what business
models should be based around.

Take for example Lego, a successful company that was near collapse in the
early 2000’s because they lost touch with their customers. After failed
attempts to reposition the company with action figures and other concepts,
Jørgen Vig Knudstorp, CEO of the Danish Lego firm, decided to engage in a
major qualitative research project. Children in five major global cities were
studied to help Lego better understand the emotional needs of children in
relation to legos. After evaluating hours of video recordings of children
playing with legos, a pattern emerged. Children were passionate about the
“play experience” and the process of playing. Rather than the instant
gratification of toys like action figures, children valued the experience of
imagining and creating. The results were clear; Lego needed to go back to
marketing its traditional building blocks and focus less on action figures and
toys. Today, Lego is once again a successful company, and thick data proved
to be its savior.

THIN DATA

It provides information or deep insights less compared to thick data.

Casestudy

Strengths

·10 Very large data set containing records of approximately 5.7% of UK


population.

·11 Broadly representative of the UK population.

·12 Allows all patients with a particular disease to be studied.

·13 Can select control subjects from the same source population.

·14 Data are collected in a non-interventional way and therefore reflect


‘real life’

·15 Information is continually updated, permitting investigation of the


effects new interventions/treatments.

·16 Data does not have to be collected which may reduce time and costs

·17 Is amenable to most epidemiological study designs (i.e. cohort, case-


control, case-series).
·18 Can be used to study relatively rare exposures or outcomes.

·19 A number of practices have been linked to HES (hospital episode


statistics) data.

Limitations

·20 Primary use of Vision software is patient management and not


medical research and data will reflect only those events that are
deemed to be relevant to the patient’s care.

·21 Analysis will require an experienced data manager with access to


appropriate computer software and hardware.

·22 Even such a large data set may have power problems when exposure
and outcomes are both rare.

·23 Situations where THIN data may not be appropriate include:

·24 Studies where individual ethnicity, occupation, employment, and/or


socio-economic status are important variables (not available as
patient level data).

·25 Non-compliance to medication prescriptions may be an issue for


drug-related exposures.

·26 Studies where data are primarily related to secondary care (cancer-
care studies).

·27 Studies examining over the counter (OTC) drugs as these will not
usually be recorded.

Note: Studies looking at lab test results before computerization should be


aware that only abnormal values may have been entered.

TRANSACTIONAL DATA

Transaction data are data describing an event (the change as a result of a


transaction) and is usually described with verbs. Transaction data always
has a time dimension, a numerical value and refers to one or more objects
(i.e. the reference data). Typical transactions are: Financial: orders, invoices,
payments.

Example: Paypal manages the largest transactional data in the world.

SPATIAL DATA

Also known as geospatial data or geographic information it is the data or


information that identifies the geographic location of features and
boundaries on Earth, such as natural or constructed features, oceans, and
more. Spatial data is usually stored as coordinates and topology, and is data
that can be mapped.

Example: Its shows that Japan's geographical status is prone to volcanic,


tsunami and many other natural attacks.

You might also like