This document discusses big data, providing examples and characteristics. It defines big data as data that is huge in volume and growing exponentially over time. Examples given include trade data from the New York Stock Exchange and data from Facebook. The characteristics of big data are described as volume, variety, velocity, and variability. Advantages of processing big data are also summarized.
This document discusses big data, providing examples and characteristics. It defines big data as data that is huge in volume and growing exponentially over time. Examples given include trade data from the New York Stock Exchange and data from Facebook. The characteristics of big data are described as volume, variety, velocity, and variability. Advantages of processing big data are also summarized.
Topic :: Big Data Submitted By:: Muhammad Saleem Math211101090 Submitted To:: Dr Saima Noreen Khosa
Dated:: 1/3/2024
Big Data is a collection of data that is huge in volume, yet growing
exponentially with time. It is a data with so large size and complexity that none of traditional data management tools can store it or process it efficiently. Big data is also a data but with huge size. Following are some of the Big Data examples, The New York Stock Exchange is an example of Big Data that generates about one terabyte of new trade data per day. Social Media The statistic shows that 500+terabytes of new data get ingested into the databases of social media site Facebook, every day. This data is mainly generated in terms of photo and video uploads, message exchanges, putting comments etc.
Types Of Big Data
1. Structured 2. Unstructured 3. Semi-structured Structured Any data that can be stored, accessed and processed in the form of fixed format is termed as a ‘structured’ data. Over the period of time, talent in computer science has achieved greater success in developing techniques for working with such kind of data (where the format is well known in advance) and also deriving value out of it. However, nowadays, we are foreseeing issues when a size of such data grows to a huge extent, typical sizes are being in the rage of multiple zettabytes. 1021 bytes = 1 zettabyte or 1 billion terabytes Examples Of Structured Data An ‘Employee’ table in a database is an example of Structured Data Unstructured Any data with unknown form or the structure is classified as unstructured data. In addition to the size being huge, un-structured data poses multiple challenges in terms of its processing for deriving value out of it. A typical example of unstructured data is a heterogeneous data source containing a combination of simple text files, images, videos etc. Now day organizations have wealth of data available with them but unfortunately, they don’t know how to derive value out of it since this data is in its raw form or unstructured format. Examples Of Un-structured Data The output returned by ‘Google Search’ Semi-structured Semi-structured data can contain both the forms of data. We can see semi- structured data as a structured in form but it is actually not defined with e.g. a table definition in relational DBMS. Examples Of Semi-structured Data Personal data stored in an XML file Data Growth over the years Please note that web application data, which is unstructured, consists of log files, transaction history files etc. OLTP systems are built to work with structured data wherein data is stored in relations (tables).
Characteristics Of Big Data
Big data can be described by the following characteristics: Volume Variety Velocity Variability (i) Volume –The name Big Data itself is related to a size which is enormous. Size of data plays a very crucial role in determining value out of data. Also, whether a particular data can actually be considered as a Big Data or not, is dependent upon the volume of data. (ii) Variety –Variety refers to heterogeneous sources and the nature of data, both structured and unstructured. During earlier days, spreadsheets and databases were the only sources of data considered by most of the applications. Nowadays, data in the form of emails, photos, videos, monitoring devices, PDFs, audio, etc. are also being considered in the analysis applications. This variety of unstructured data poses certain issues for storage, mining and analyzing data. (iii) Velocity –The term ‘velocity’ refers to the speed of generation of data. How fast the data is generated and processed to meet the demands, determines real potential in the data. Big Data Velocity deals with the speed at which data flows in from sources like business processes, application logs, networks, and social media sites, sensors,Mobile devices, etc. (iv) Variability –This refers to the inconsistency which can be shown by the data at times, thus hampering the process of being able to handle and manage the data effectively. Advantages Of Big Data Processing Ability to process Big Data in DBMS brings in multiple benefits, such as- Businesses can utilize outside intelligence while taking decisions Access to social data from search engines and sites like Facebook, Twitter are enabling organizations to fine tune their business strategies. Improved customer service Traditional customer feedback systems are getting replaced by new systems designed with Big Data technologies. In these new systems, Big Data and natural language processing technologies are being used to read and evaluate consumer responses. Early identification of risk to the product/services, if any Better operational efficiency Big Data technologies can be used for creating a staging area or landing zone for new data before identifying what data should be moved to the data warehouse. In addition, such integration of Big Data technologies and data warehouse helps an organization to offload infrequently accessed data.