ICT703_Big Data_Assessment 1_ Case Study Analysis Report_1.2 (1)
ICT703_Big Data_Assessment 1_ Case Study Analysis Report_1.2 (1)
Document Details
Submission ID
trn:oid:::3618:74026999 11 Pages
Download Date
File Name
File Size
496.6 KB
The percentage indicates the combined amount of likely AI-generated text as It is essential to understand the limitations of AI detection before making decisions
well as likely AI-generated text that was also likely AI-paraphrased. about a student’s work. We encourage you to learn more about Turnitin’s AI detection
capabilities before using the tool.
Detection Groups
1 AI-generated only 0%
Likely AI-generated text from a large-language model.
Disclaimer
Our AI writing assessment is designed to help educators identify text that might be prepared by a generative AI tool. Our AI writing assessment may not always be accurate (it may misidentify
writing that is likely AI generated as AI generated and AI paraphrased or likely AI generated and AI paraphrased writing as only AI generated) so it should not be used as the sole basis for
adverse actions against a student. It takes further scrutiny and human judgment in conjunction with an organization's application of its specific academic policies to determine whether any
academic misconduct has occurred.
False positives (incorrectly flagging human-written text as AI-generated) are a possibility in AI models.
AI detection scores under 20%, which we do not surface in new reports, have a higher likelihood of false positives. To reduce the
likelihood of misinterpretation, no score or highlights are attributed and are indicated with an asterisk in the report (*%).
The AI writing percentage should not be the sole basis to determine whether misconduct has occurred. The reviewer/instructor
should use the percentage as a means to start a formative conversation with their student and/or use it to examine the submitted
assignment in accordance with their school's policies.
Non-qualifying text, such as bullet points, annotated bibliographies, etc., will not be processed and can create disparity between the submission highlights and the
percentage shown.
Assessment 1
Cloud-Based Big Data Solutions: AWS, Azure,
and Google Cloud
SUBMITTED BY:
Table of Contents
1. Abstract 3
2. Introduction 3
3. Contemporary Issues in Big Data 4
4. Current State of Knowledge 5
5. Critical Evaluation of Technologies 6
6. Impact on Business and Society 10
7. Future Trends and Recommendations 11
8. Conclusion 11
References 11
1. Abstract
Statistical data processing is rapidly changing as organisations’ approach to managing data for purposes
of decision making. In this report, the author examines various peculiarities and development trends of
big data systems engineering [1]. In the systematic literature review conducted in the analysis, the key
characteristics of scalability, processing, and security are presented; and important frameworks such as
Hadoop and Spark are discussed. In the present work, I have critically discussed these technologies and
have underlined how such technologies are reshaping business. Finally, some recommendations for
further study and actual development of big data systems are provided.
2. Introduction
1. Background
The term big data is defined as large amounts of structured and unstructured data collected at high speeds.
(Amazon SageMaker) When organizations need to make decisions that are based on data, there are
questions like how to address scalability, security, or updating information in real-time. Some of these
challenges have been overcome by systems such as Apache Hadoop and Apache Spark but there are still
shortcomings in the adaptability and optimization of these systems.
2. Relevance
Big data has now become a must-have and not a nice to have for any business competing in today’s
market. Starting from the consumer market to the healthcare domain, (Gandomi and Haider 2015), big
data is used to forecast patterns, analyze performance, and tailor experiences. Information systems
professionals should have adequate knowledge concerning these technologies, so they may handle the
associated problems and optimize the technologies’ value.
3. Objective
The purpose of this report is to discuss modern challenges in big data design, assess the current
technologies, and consider prospects for further development. (G. N. Okorie , 2024), It endeavours to do
so by identifying existing shortcomings and innovative phenomena in order to help professionals and
researchers refine their systems.
4. Scope
The fast-growing data adds scalability issues on the current systems. (A. Martynenko 2020), Clients need
expanding volumes to be addressed as well as perseverance of the first message without losing time. Real
time processing is another key area, especially for items such as fraud detection services and
recommendation engines. Protection needs, especially for information assets, are increasing due to
emerging cybersecurity risks. Such issues call for creative steps in framework and architecture towards
building sound and effective systems.
Figure 1: Big Data Processing Pipeline: From Data Ingestion to Insight (Self Drawn)
The latest literature review focuses on the emerging big data frameworks. (J. T. Hancock 2020), Most
major issues of scalability were tackled during the improvement of the first systems; Starting with
Hadoop, distributed storage (HDFS) and processing (MapReduce) technologies were used. Apache Spark
that enables in-memory computation is credited for transforming real-time processing. AWS, Azure and
others contain advantages of a decentralized structure and including scalability, available on demand
resources. However, concerns like integration and cost effectiveness still fall in the research frontier.
Emerging Big data technology is finding lot many use cases in various industries. The application is
innovative and helping business accomplish their objectives. A set of use cases are shown in figure 3.
Advantages:
● Rich set of services such as EMR (Elastic MapReduce) used for big data processing.
● Heightened availability attributable to the world's data centers.
● Integrated with the AWS service SageMaker for machine learning.
Limitations:
● High and increasing costs for data storage (P. Maroufkhani 2023) and computation beyond the
immediate needs in topical time.
● This is due to the risk of locking in with the specific vendor propelling the services used as
proprietary.
Use Cases:
● Recommendation IoT data analytics, real-time log analysis recommendation systems.
2. Microsoft Azure
Advantages:
● System integration with Microsoft products (Microsoft Azure), such as the Power Business
Intelligence application, the SQL Server.
● Azure Synapse has capabilities to support big data as well as a traditional data warehouse.
● Delivers both for with and without installation in a client’s infrastructure.
Limitations:
● Complex pricing model.
● Possesses somewhat less compatibility with tools outside the Microsoft ecosystem than
competitors.
Use Cases:
● Predict maintenance, operation analytics, and hybrid cloud solutions.
Advantages:
● BigQuery stability and TensorFlow as well as other strong AI & Machine Learning tools
available in the cloud.
● Better time-driven processing through Dataflow and Pub/Sub.
● Low cost offering with per second metering for computer instances.
Limitations:
● Even if a smaller player compared to AWS or Azure it has a more exposed global presence.
● Third party integration services tend to be fewer in number.
Use Cases:
● Real time customer analytics, the integration, training and deployment of a machine learning
model and processing genomic data.
I have tried to do comparison between three super scalars in terms of the features and the same has been
presented in table 1 below.
Figure 5: Performance Comparison Chart of AWS, Microsoft Azure and Google Cloud Platform
Analytics on volume of data results in increased productivity, effective promotion and selling strategies
and improved business strategies and tactics. For instance, (Fan, Han, and Liu 2014) Amazon relies on
big data to make suggestions on the products to offer their customers and healthcare providers use big
data to determine the possible future health of patients. Nonetheless, keeping data private becomes an
ethical issue because they need more control and clear policies for using them.
Figure 6: Scalability and Performance Comparison of Big Data Platforms (Self Drawn)
Big data is and will continue to be incorporating artificial intelligence as a means of performing predictive
analytics, as well as automating decision making systems. (Guo and Chen 2023), The Big data technology
could help to solve the problems of security and transparency. Firms should implement hybrid solutions
that will include the use of both, on-premise and Cloud systems in order to cut on expenses while at the
same time increasing efficiency. Based on research, there is the need to develop easy to implement and
integrate frameworks that can reduce entry barriers to new business ventures.
8. Conclusion
This report highlights for the importance of big data in contemporary business intelligence. This is
because frameworks such as Hadoop and Spark eases issues of scalability and real time data processing,
but cost effectiveness and security still present challenges. All the recommendations given here are
intended to help business to design and establish more solid big data systems. More investigation needs to
go towards blending the technology and data sources to increase the outcome performance and security.
References