UNIT 1-Capstone Project Practice Questions
UNIT 1-Capstone Project Practice Questions
9. Which concept involves the breakdown of time series data into trend, seasonality, and noise
components?
• A) Time Series Forecasting
• B) Design Thinking
• C) Problem Decomposition
• D) Time Series Decomposition
12. Which of these approaches would you use for showing relationships between variables?
• A) Predictive approach
• B) Descriptive approach
• C) Classification approach
• D) Regression
17. Which concept refers to adjusting models using new data to improve their accuracy?
• A) Refinement
• B) Validation
• C) Cross-validation
• D) Feature selection
19. What percentage is commonly used for training data in a train-test split?
• A) 80%
• B) 20%
• C) 67%
• D) 50%
20. In a cross-validation process, how many subsets are generally created in a 5-fold cross-
validation?
• A) 2
• B) 5
• C) 10
• D) 3
21. When is cross-validation more beneficial than train-test split?
• A) For large datasets
• B) For datasets with limited rows
• C) When doing unsupervised learning
• D) For high computational costs
22. Which of the following is a commonly used metric for regression models?
• A) Accuracy
• B) Precision
• C) Recall
• D) Root Mean Squared Error (RMSE)
26. Which algorithm is used in the example of the Airline Passenger Dataset?
• A) Decision Tree
• B) Random Forest
• C) Seasonal Decomposition
• D) Support Vector Machine
30. In a recommendation system, which method is typically used to suggest new items?
• A) Clustering
• B) Regression
• C) Collaborative filtering
• D) Anomaly detection
32. Which is the most reliable method to evaluate model performance on smaller datasets?
• A) Simple train-test split
• B) Leave-one-out cross-validation
• C) Randomized testing
• D) Bootstrap aggregation
39. Which error metric penalizes large errors more than small errors?
• A) RMSE
• B) MSE
• C) Accuracy
• D) Precision
40. Which error metric would you use to compare different regression models?
• A) Classification accuracy
• B) RMSE
• C) ROC-AUC score
• D) F1-Score
41. Which of the following is most important when evaluating a model’s accuracy on unseen
data?
• A) Precision
• B) Validation data
• C) Recall
• D) Feature engineering
42. Which metric is used to evaluate classification tasks in binary classification?
• A) Precision and recall
• B) RMSE
• C) MSE
• D) MAE
43. Which evaluation metric balances precision and recall in a classification problem?
• A) F1-Score
• B) Accuracy
• C) RMSE
• D) Cross-validation
46. Which evaluation metric is best for highly imbalanced classification datasets?
• A) Accuracy
• B) F1-Score
• C) RMSE
• D) MAE
47. Which AI project involves recognizing human activities using smartphone data?
• A) Stock Prices Predictor
• B) Human Activity Recognition
• C) Student Results Predictor
• D) Sentiment Analysis
1. Assertion (A): The Capstone Project integrates all learning from an academic program.
Reason (R): It focuses solely on individual work rather than collaboration.
4. Assertion (A): Data scientists use training sets to evaluate model performance.
Reason (R): Test sets are used to adjust models after training is complete.
5. Assertion (A): Cross-validation ensures more reliable model evaluation than a train-test
split.
Reason (R): Cross-validation evaluates models using different data folds, making the process
more computationally efficient.
6. Assertion (A): RMSE (Root Mean Squared Error) is a commonly used metric for
evaluating regression models.
Reason (R): RMSE penalizes larger errors more significantly than smaller errors, making it
sensitive to outliers.
8. Assertion (A): MSE (Mean Squared Error) penalizes large errors more severely than
RMSE.
Reason (R): MSE focuses on the average squared difference between predicted and actual
values.
A) Both A and R are true, and R is the correct explanation of A.
B) Both A and R are true, but R is not the correct explanation of A.
C) A is true, but R is false.
D) A is false, but R is true.
2. A student team is tasked with building a sentiment analyzer that classifies text as positive,
negative, or neutral. They collected a dataset of tweets but discovered that the data includes
irrelevant information like URLs and emojis.
Question: How should the team handle the data before building their AI model?
3. A team is working on a project to address the issue of crop yield prediction in agriculture.
They collected a large dataset but are unsure which AI model to use for this type of prediction.
Question: What type of model should the team consider for predicting crop yields, and why?
4. A group is predicting brain weights based on head size using linear regression. After running
their model, they calculated an RMSE of 73.
Question: How should the team interpret this RMSE value, and what should be their next step?
6. A team is working on predicting movie ticket prices based on factors such as location, movie
type, and time of day. They have used a dataset but are not sure if their model is performing well.
Question: What metrics should the team use to evaluate their model’s performance?
7. A team is tasked with using AI to predict patient recovery times based on medical history and
treatment data. However, the data has missing values.
Question: What steps should the team take to handle the missing data before applying their AI
model?
QUESTION-ANSWERS:
1. What is a Capstone Project in the context of AI education?
19. Why is the business understanding stage crucial in data science projects?
21. How does the “empathize” stage in Design Thinking help in AI projects?
22. What is the purpose of using a prototype in the Design Thinking process?
8. Imagine that you want to create your first app. This is a complex problem. How would
you decompose the task of creating an app?
12. What type of Questions can be asked for the following approaches:
16. What are the necessary things to be done for the success of Data modelling stage?
21. What are the prerequisites for Train and Test Data?
26. Study the following graph and answer the given questions: