CS614 Midterm Grand Solved Quiz
CS614 Midterm Grand Solved Quiz
VULMSHELP.COM
JUNAID MALIK
0304-1659294
AL-JUNAID INSTITUTE GROUP
CS614 GRAND QUIZ
Onlline Extraction is a kind of---------------------------------data extraction.
o Logical
o Dimensional
In context of data warehouse, normally it becomes difficult to extract data from different
sources because these sources are normally.
• Heterogeneous page 140
• Homogeneous
• Centralized
• Baseline
Which people criticize Dimensional Modeling (DM) as being a data mart oriented approach?
• Those that consider ER models as Data marts
• Those that consider Business processes as Data marts page 110
• Those that consider Data marts as Data warehouse
• Those that consider dimensional model
• Those that consider dimensional modeling as de-normalization approach
The data has to be checked , cleaned and transformed into a --------------- format to allow easy
and fast access
• Unified page 20
• Predicated
• Qualified
• Proactive
Suppose in a system A, the values of “PhoneNo” attribute were stored in “countrycode-phone-
extension” format, however after transformation into data warehouse the separate columns
were used for “countrycode”,”phone” and “extension”. The above scenario is an example of :
• One-to-one scalar transformation
• One-to-many element transformation page 144+conceptual
• Many-to-one element transformation
• Many-to-many element transformation
In decision support system ease of use in achieved by:
• Normalization
• Denormalization page no 49
• Drill up
• Drill down
Which of the following is one of the methods to simplify an ER model?
• Normalization
• Denormalization page no 103
• HOLAP
• Hybrid schema
In ETL process data transformation includes ----------------
AL-JUNAID INSTITUTE GROUP
• Data cleansing page 129
• Data aggregation
• Behavior checking
• Pattern recognition
Non-uniform use of abbreviations, units, and values refers to:
• Syntactically dirty data page 160
• HOLAP
• DOLAP page78
• ROLAP
• MOLAP
The main reason(s )for the increase in cube size may be:
• Increase in the number of dimensions
• Increase in the cardinality of the dimensions
• Increase in the amount of detail data
• All of the given options page 87
Suppose the amount of data recorded in an organization is doubled in year. This increase in ------
• Linear
• Quadratic
• Exponential page 15
• Logarithmic
The data in the data warehouse is ----------- :
• Volatile
AL-JUNAID INSTITUTE GROUP
• Non-volatile page 69
• Static
• Non-structured
--------------- models the macro relationships among data elements with an overall deterministic
strategy:
• Dimensional model page102
• Entity relationship model
• Object oriented model
• Structured model
----------------- technique requires a separate column to specify the time and date when the last
modification was occurred:
• Checkmarks
• Timestamps page 150
• Just-in-Time
• Real Time extraction
Which of the de-normalization technique squeezes master table into detail?
• Pre-joining page 58
• Horizontal splitting
• Vertical splitting
• Adding redundant column
De-normalization can help:
• Minimize joins
• Minimize foreign keys
• Resolve aggregates
• All of the given optionspage 51
The domain of the “gender” field in some database may be (‘F’,’M’) or as (“Female”, “Male”)
or even as (1, 0). This is:
• Primary key problem
• Non primary key problem page 163
• Normalization problem
• All of the given option
Increasing level of normalization number of
---------------- tables:
• Increases page 51
• Decreases
• Does not effect
AL-JUNAID INSTITUTE GROUP
• None of the given option
Which of the following is not a Data Quality Validation Technique:
• Referential integrity
• Using Data Quality Rules
• Data Histograming
• Indexes page 189
This technique can be used when column from one table is frequently accessed in a large scale
join in conjunction with a column from another table:
• Horizontal splitting
• Pre-joining
• Adding redundant column page 58
• Derived attributes
Data cleansing requires involvement of domain expert because:
• Domain expert has deep knowledge of data aggregation
• Change Data captures requires involvement of domain expert
• Domain knowledge is required to correct anomalies page 158
• Domain expert has deep knowledge of data summarization
Relational databases allow you to navigate the data in ------------- that is appropriate using the
primary , foreign key structure with in the data model:
• Only One Direction
• Any Direction page 19
• Two Direction
• None of these
History is excellent predicator of the ------------:
• Past
• Present
• Future page 15
• History
De- normalization is the process of selectively transforming normalized relations into un-
normalized physical record specifications, with the aim to:
---------------- Breaks a table into multiple tables based upon common column values
Horizontal page
• splitting 54
• Vertical splitting
• Adding redundant column
• None of the given option
Which of the following is NOT an example of derived attribute?
• Age
• CGPA
• Area of rectangle
• Height (Conceptual)
The growth of master files and magnetic tapes exploded around the mid- ---------------
• 1950s
• 1960s page 12
• 1970s
• 1980s
If one or more records in a relational table do not satisfy one or more integrity constraint , then
the data:
• Is syntactically dirty
• Is semantically dirty page 160
• Has Coverage anomaly
• Has extraction issue
OLAP is:
• Analytical processing page 69
• Transaction processing
• Additive processing
• Active processing
One of the possible issues faced by web scrapping is that:
• Web pages may contain junk data page 141
• Web pages do not contain multiple facts
• Web pages do not contain multiple dimensions
• Web pages does not support transformation
In case of multiple sources for the same data element , we need to prioritize the source systems
per element based, the process is called:
• Ranking page 143
• Prioritization
• Element selection
• Measurement selection
One feature of Change Data Capture (CDC) is that:
• It pre-calculates changed aggregates
• It loads the transformed data in real time
• It only processes the data has been changed
• It can automate the transformation of extracted data page 149
In ------------------ SQL generation in vastly simplified for front-end tools when the data is
highly structure:
• MOLAP
• Star Schema page 107
• Hybrid schema
• Object oriented schema
Dirty data means:
• Data cannot be aggregated
• Data contains non-additive facts
• Data does not fulfill dimensional modeling rules
• Data does not conform to proper domain definitions page 158
In Context of Change Data Capture (CDC) sometimes a ------------- object can be used to
store recently modified data:
• Buffer table
• Change table page 149
• Checkmark table
• Change control table
AL-JUNAID INSTITUTE GROUP
“Sometimes during data collection complete entities are missed”. This statement is an example
of :
• Missing tuple page 161
• Missing attribute
• Missing aggregates
• Semantically dirty data
Table collapsing technique is applied in case of:
• One-by-one relation or many-to –many relation page 52
• One-to-many relation
• Many-to-many relation
• None of the given option
Which of the following is an example of dimension?
• Product
• Region
• Date
• All of the given option page 78
Data warehouse stores -------------------:
• Operational data
• Historical data page 24
• Meta data
• Log files data
The business process covered by ER diagrams:
• Do not co-exist in time and space page 109
• Co-exist in time and space
• Do not physically exist in real time context
• None of the given options
The main goal of normalization is to eliminate -----------:
• Data redundancy page 41
• Data sharing
• Data security
• Data consistency
Serious ---- involves decomposing and resembling the data:
• Data cleansing page 168
• Data transformation
• Data loading
• Data extraction
AL-JUNAID INSTITUTE GROUP
In the data warehouse environment the data is ------------
• Subject- oriented page 69
• Time- oriented
• Both subject and time oriented
• Neither time-oriented nor subject- oriented
For large record spaces and large number of records , the run time of the clustering
algorithms:
• Prohibitive page 164
• Static
• Exponential
• Numerical
------------- can result in costly errors, such as , False frequency distributions and incorrect
aggregates due to double counting:
• Data duplication page 165
• Data reduction
• Data anomaly
• Data transformation
The degree to which values are present in the attributes that require them is known as -
---------------------:
• Completeness page 185
• Uniqueness
• Accessibility
• Consistency
Time complexity of Key Creation process in basic Sorted Neighborhood (BSN) Method is
----------------------:
• O(n log n)
• O(log n)
• O(n) page 171
• O(2n)
Which of the following is an example of slowly changing dimensions?
• Inheritance page 124
• Aggregation
• Association
• Asset disposal
The ------------ operator proves useful in more complex metrices applicable to the
dimensions and accessibility:
• Max page 188
AL-JUNAID INSTITUTE GROUP
• Min
• Max and Min
• None of the given
In OLAP , the typical write operation is ------------- :
• Bulk insertion page 75
• Single insertion
• Sequential insertion
• No insertion
The issue(s) of “ Adding redundant column” includes(s):
• Increase in table size
• Maintenance
• Loss of information
• All of the given option page 65
Simple scalar transformation is a--------------mapping from one set of values to another set of
values using straightforward rules.
• One-to-one page 144
• One-to-many
• Many-to-many
• Many-to-one
---------------can be created in operational systems to keep tracks of recently updated
records.
• Triggers page 150
• Timestamps
• Partitioning
• ELT
Development of data warehouse is hard because data sources are usually--------
• Structured and homogeneous
• Unstructured and heterogeneous page 31
•• Structured andand
Unstructured heterogeneous
homogeneous
In a decision support environment, the decision maker is interested in -------------.
• Only limited organizational data
• Big picture of organizational data page 21
• Only sale related data
• Only customer related data
Information can answer question like “what”, “who” and “when” while knowledge
can answer question like----------------.
• Why
• Where
• Which
• How page 11
OLTP implementations are fully-------------.
• Normalized page 69
• Denormalized
• Predictive
• Additive
Which logical data extraction has significant performance impacts on the data warehouse
server?
AL-JUNAID INSTITUTE GROUP
• Incremental Extraction page 133
• Online Extraction
• Offline Extraction
• Legacy Vs OLTP
Consider the following Student table and identify the column which causes that the table is
not in first normal form(1NF).
Student(Std_ID, Std_Name ,Std_CGPA ,Std_Hobbies)
• Std_ID
• Std_Name
• Std_CGPA
• Std_Hobbies page 43(Conceptual)
Analytical processing uses ---------------
• Multi-level aggregates page 74
• Record level aggregates
• Table level aggregates
• All of the given options
Which is not a class of anomalies in following?
• Dirty anomalies page 160
• Syntactically dirty data
• Semantically dirty data
• Coverage anomalies
Which one among the following data warehouse stores data containing long period?
• Telecommunication data warehouse
• Financial data warehouse
• Human resource data warehouse
• Insurance data warehouse page 36