0% found this document useful (0 votes)
53 views

UNIT 1-Capstone Project Practice Questions

Capstone project

Uploaded by

Rashmi Kaith
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
53 views

UNIT 1-Capstone Project Practice Questions

Capstone project

Uploaded by

Rashmi Kaith
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 14

UNIT 1: CAPSTONE PROJECT

1. What is the main purpose of a Capstone Project?


o A) To demonstrate theoretical knowledge.
o B) To complete a thesis paper.
o C) To integrate all knowledge gained through a comprehensive project.
o D) To learn about different industries.

2. Which of the following is not an objective of a Capstone Project?


o A) Solving real-world problems.
o B) Expressing solutions in technical terms.
o C) Selecting appropriate algorithms for a problem.
o D) Learning teamwork.

3. Which AI project involves predicting stock prices?


o A) Movie Ticket Price Predictor
o B) Stock Prices Predictor
o C) Sentiment Analyzer
o D) Student Results Predictor

4. Which AI model is typically used for classification?


o A) Regression
o B) Clustering
o C) Classification
o D) Anomaly Detection

5. What is the first step in the AI project cycle?


o A) Model construction
o B) Data gathering
o C) Problem definition
o D) Evaluation & refinements

6. Which step is critical in determining whether AI techniques are applicable to a problem?


o A) Gathering data
o B) Identifying a pattern in data
o C) Deploying the model
o D) Selecting the right algorithm
7. Which of the following is not a stage of Design Thinking?
o A) Empathize
o B) Define
o C) Deploy
o D) Prototype

8. Which is an example of problem decomposition?


o A) Gathering data from sensors
o B) Breaking down app development into multiple tasks
o C) Running machine learning models
o D) Collecting user feedback

9. Which concept involves the breakdown of time series data into trend, seasonality, and noise
components?
• A) Time Series Forecasting
• B) Design Thinking
• C) Problem Decomposition
• D) Time Series Decomposition

10. Which is the first step in any AI or machine learning project?


• A) Data modeling
• B) Data collection
• C) Business understanding
• D) Cross-validation

11. Which is the foundational methodology for data science?


• A) Data mining
• B) CRISP-DM
• C) Agile
• D) SDLC

12. Which of these approaches would you use for showing relationships between variables?
• A) Predictive approach
• B) Descriptive approach
• C) Classification approach
• D) Regression

13. When might a predictive model be used?


• A) To explain historical data
• B) To show relationships between data
• C) To predict future outcomes
• D) To cluster similar data points
14. What question should be asked first in a data project?
• A) What is the business outcome?
• B) How will the data be collected?
• C) What data is needed?
• D) What algorithm will be used?

15. Which dataset is commonly used for predicting house prices?


• A) Airline Passenger Dataset
• B) Forestfires Dataset
• C) Housing Dataset
• D) MNIST Dataset

16. What is a descriptive model used for?


• A) Prediction of new data
• B) Describing relationships in historical data
• C) Anomaly detection
• D) Identifying missing data

17. Which concept refers to adjusting models using new data to improve their accuracy?
• A) Refinement
• B) Validation
• C) Cross-validation
• D) Feature selection

18. What does the train-test split method achieve?


• A) Collecting the data
• B) Evaluating model performance
• C) Data pre-processing
• D) Model deployment

19. What percentage is commonly used for training data in a train-test split?
• A) 80%
• B) 20%
• C) 67%
• D) 50%

20. In a cross-validation process, how many subsets are generally created in a 5-fold cross-
validation?
• A) 2
• B) 5
• C) 10
• D) 3
21. When is cross-validation more beneficial than train-test split?
• A) For large datasets
• B) For datasets with limited rows
• C) When doing unsupervised learning
• D) For high computational costs

22. Which of the following is a commonly used metric for regression models?
• A) Accuracy
• B) Precision
• C) Recall
• D) Root Mean Squared Error (RMSE)

23. Which metric is most suitable for classification tasks?


• A) MSE
• B) Accuracy
• C) RMSE
• D) Noise ratio

24. Which of the following is used to calculate RMSE?


• A) Mean of residuals
• B) Sum of absolute errors
• C) Square root of the mean of squared errors
• D) Mean of absolute differences

25. What does a low RMSE indicate?


• A) Poor model performance
• B) High variance in predictions
• C) Accurate predictions
• D) Overfitting

26. Which algorithm is used in the example of the Airline Passenger Dataset?
• A) Decision Tree
• B) Random Forest
• C) Seasonal Decomposition
• D) Support Vector Machine

27. Which value represents the best prediction in MSE?


• A) The highest value
• B) The lowest value
• C) The mean of predictions
• D) The median of predictions
28. What type of learning involves algorithms like regression or classification?
• A) Supervised learning
• B) Unsupervised learning
• C) Reinforcement learning
• D) Semi-supervised learning

29. Which algorithm is most suitable for a regression problem?


• A) Decision tree
• B) Linear regression
• C) K-nearest neighbors
• D) Naive Bayes

30. In a recommendation system, which method is typically used to suggest new items?
• A) Clustering
• B) Regression
• C) Collaborative filtering
• D) Anomaly detection

31. Which of the following would be considered a feature in a dataset?


• A) The target label
• B) An algorithm
• C) A variable used for prediction
• D) The test set

32. Which is the most reliable method to evaluate model performance on smaller datasets?
• A) Simple train-test split
• B) Leave-one-out cross-validation
• C) Randomized testing
• D) Bootstrap aggregation

33. Cross-validation is typically used to:


• A) Build the model
• B) Split data into train and test sets
• C) Test the model with multiple subsets
• D) Apply unsupervised learning

34. What is one major drawback of cross-validation compared to train-test split?


• A) It uses less data.
• B) It takes more time and computational resources.
• C) It produces less accurate results.
• D) It can only be applied to classification problems.
35. Which validation method involves using every data point for testing at least once?
• A) K-fold cross-validation
• B) Simple validation
• C) Random split
• D) Hold-out validation

36. What is the primary advantage of using cross-validation?


• A) Requires fewer computational resources
• B) More accurate representation of model performance
• C) Faster training of the model
• D) Higher accuracy for large datasets

37. What is the goal of hyperparameter tuning?


• A) Choosing the right model
• B) Optimizing algorithm performance
• C) Collecting more data
• D) Scaling the data

38. What does MAPE stand for?


• A) Mean Absolute Prediction Error
• B) Mean Absolute Percentage Error
• C) Mean Adjusted Prediction Error
• D) Minimum Absolute Prediction Estimate

39. Which error metric penalizes large errors more than small errors?
• A) RMSE
• B) MSE
• C) Accuracy
• D) Precision

40. Which error metric would you use to compare different regression models?
• A) Classification accuracy
• B) RMSE
• C) ROC-AUC score
• D) F1-Score

41. Which of the following is most important when evaluating a model’s accuracy on unseen
data?
• A) Precision
• B) Validation data
• C) Recall
• D) Feature engineering
42. Which metric is used to evaluate classification tasks in binary classification?
• A) Precision and recall
• B) RMSE
• C) MSE
• D) MAE

43. Which evaluation metric balances precision and recall in a classification problem?
• A) F1-Score
• B) Accuracy
• C) RMSE
• D) Cross-validation

44. What does MAE stand for in machine learning?


• A) Model Accuracy Estimate
• B) Mean Absolute Error
• C) Maximum Accuracy Estimate
• D) Minimum Adjustment Error

45. Which error metric is less sensitive to outliers in regression problems?


• A) RMSE
• B) MAE
• C) MSE
• D) Cross-entropy

46. Which evaluation metric is best for highly imbalanced classification datasets?
• A) Accuracy
• B) F1-Score
• C) RMSE
• D) MAE

47. Which AI project involves recognizing human activities using smartphone data?
• A) Stock Prices Predictor
• B) Human Activity Recognition
• C) Student Results Predictor
• D) Sentiment Analysis

48. Which of the following best describes anomaly detection?


• A) Grouping similar data points
• B) Identifying unusual patterns in data
• C) Predicting continuous outcomes
• D) Labeling data based on features
49. In AI, what is a common use of clustering algorithms?
• A) Predicting future outcomes
• B) Grouping similar data points without labels
• C) Detecting anomalies
• D) Improving model accuracy

1. Assertion (A): The Capstone Project integrates all learning from an academic program.
Reason (R): It focuses solely on individual work rather than collaboration.

 A) Both A and R are true, and R is the correct explanation of A.


 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

2. Assertion (A): Data gathering is a critical step in an AI project cycle.


Reason (R): Without proper data, the AI model cannot be trained effectively.

 A) Both A and R are true, and R is the correct explanation of A.


 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

3. Assertion (A): AI development is always suitable for every type of problem.


Reason (R): AI techniques are applied when a pattern exists in the data.

 A) Both A and R are true, and R is the correct explanation of A.


 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

4. Assertion (A): Data scientists use training sets to evaluate model performance.
Reason (R): Test sets are used to adjust models after training is complete.

 A) Both A and R are true, and R is the correct explanation of A.


 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

5. Assertion (A): Cross-validation ensures more reliable model evaluation than a train-test
split.
Reason (R): Cross-validation evaluates models using different data folds, making the process
more computationally efficient.

 A) Both A and R are true, and R is the correct explanation of A.


 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

6. Assertion (A): RMSE (Root Mean Squared Error) is a commonly used metric for
evaluating regression models.
Reason (R): RMSE penalizes larger errors more significantly than smaller errors, making it
sensitive to outliers.

 A) Both A and R are true, and R is the correct explanation of A.


 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

7. Assertion (A): The final stage in model evaluation is deployment.


Reason (R): Deployment occurs after thorough testing, validation, and refinement of the model.

 A) Both A and R are true, and R is the correct explanation of A.


 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

8. Assertion (A): MSE (Mean Squared Error) penalizes large errors more severely than
RMSE.
Reason (R): MSE focuses on the average squared difference between predicted and actual
values.
 A) Both A and R are true, and R is the correct explanation of A.
 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

9. . Assertion (A): Data preprocessing is not necessary if the dataset is large.


Reason (R): Large datasets inherently contain all necessary information and require no
adjustments.

 A) Both A and R are true, and R is the correct explanation of A.


 B) Both A and R are true, but R is not the correct explanation of A.
 C) A is true, but R is false.
 D) A is false, but R is true.

CASE STUDY BASED QUESTIONS:


1. A team of students is working on a stock price prediction model as part of their Capstone
Project. They are facing issues because the stock prices show a lot of volatility, and the
patterns are not clear. The team is unsure about how to proceed with building their AI
model.
Question: What should the team do first before applying any AI model?

2. A student team is tasked with building a sentiment analyzer that classifies text as positive,
negative, or neutral. They collected a dataset of tweets but discovered that the data includes
irrelevant information like URLs and emojis.
Question: How should the team handle the data before building their AI model?

3. A team is working on a project to address the issue of crop yield prediction in agriculture.
They collected a large dataset but are unsure which AI model to use for this type of prediction.
Question: What type of model should the team consider for predicting crop yields, and why?

4. A group is predicting brain weights based on head size using linear regression. After running
their model, they calculated an RMSE of 73.
Question: How should the team interpret this RMSE value, and what should be their next step?

5. A student team is developing a recommendation system for improving educational resources


in schools. They want to recommend learning materials based on students’ learning habits.
Question: Which AI technique should the team use to build the recommendation system, and
why?

6. A team is working on predicting movie ticket prices based on factors such as location, movie
type, and time of day. They have used a dataset but are not sure if their model is performing well.
Question: What metrics should the team use to evaluate their model’s performance?

7. A team is tasked with using AI to predict patient recovery times based on medical history and
treatment data. However, the data has missing values.
Question: What steps should the team take to handle the missing data before applying their AI
model?

8. A group of students is working on a project to classify human activities using smartphone


sensors (like accelerometers). The data includes features such as time, accelerometer readings,
and gyroscope readings.
Question: What type of AI model should the team use for this classification task?

QUESTION-ANSWERS:
1. What is a Capstone Project in the context of AI education?

2. What are the key steps in the AI Project Cycle?

3. Why is “problem definition” important in an AI project?

4. What is Design Thinking in AI problem-solving?

5. What is time series decomposition?

6. What is the main advantage of problem decomposition in computational tasks?

7. Why is data gathering essential in an AI project?

8. What is RMSE, and why is it important in AI models?

9. What is cross-validation, and how does it improve model performance evaluation?

10. What is the purpose of a recommendation model in AI?


11. What are the key components of time series data?

12. What is the goal of AI model construction?

13. What is the difference between regression and classification in AI?

14. How is data preprocessed for AI projects?

15. What is MSE, and how is it different from RMSE?

16. Why is model validation important in AI?

17. What is the purpose of using a training dataset in AI?

18. What is an anomaly detection model used for in AI?

19. Why is the business understanding stage crucial in data science projects?

20. What is the significance of model deployment in an AI project?

21. How does the “empathize” stage in Design Thinking help in AI projects?

22. What is the purpose of using a prototype in the Design Thinking process?

23. How is the concept of clustering applied in AI?

24. What is the primary objective of feature engineering in AI?

25. Why is it important to avoid overfitting in AI models?

26. What is the role of gradient descent in machine learning?

QUESTION AND ANSWERS:


1. Define Capstone Project.

2. Give examples of Capstone Projects.

3. What are the different steps that the AI Project follows?

4. Which five types of questions should be answered during Understanding the AI


problem?

5. Define Design Thinking.

6. What are the five stages of Design Thinking?

7.What are the steps of Problem Decomposition?


Or
How to down the problem into smaller units before coding?

8. Imagine that you want to create your first app. This is a complex problem. How would
you decompose the task of creating an app?

9. Explain time series decomposition.

10. What are the components of Time series decomposition?

11. What are different Analytic approaches?

12. What type of Questions can be asked for the following approaches:

13. Explain Data Gathering Phase.


OR
How Data is collected for AI Projects?

14. Explain Data Modelling.


OR
Differentiate between Predictive and Descriptive model.

15. Define training set.

16. What are the necessary things to be done for the success of Data modelling stage?

17. Differentiate between Cross-Validation and Train-Test Split


18. Explain Train-Test Split Evaluation

19. How to Configure the Train-Test Split?


20. How to choose a split percentage in Train-Test Split?
.

21. What are the prerequisites for Train and Test Data?

22. Explain Cross validation

23. Explain loss functions.

24. What are the categories of Loss Functions?

25. Differentiate Classification Loss and Regression Loss.

26. Study the following graph and answer the given questions:

i. What does red dots represent?


ii. What does blue line represent?
iii. What does line X represents?
iv. How can we calculate RMSE using graph?
v. What is a good model based on RMSE?

27. Explain MSE.


28. When should we use mean squared error?

29. Can MSE be a negative value? Give reasons.

30. Explain RMSE

You might also like