0% found this document useful (0 votes)
4 views18 pages

50MCQ_Lecture1[1]

The document consists of a series of multiple-choice questions (MCQs) focused on word embeddings and their applications in natural language processing (NLP). Key topics include the goals of word embeddings, the advantages over one-hot vectors, various embedding methods, and the roles of different components in the Continuous Bag of Words (CBOW) model. It also addresses challenges in training and using word embeddings for tasks like sentiment analysis and machine translation.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
4 views18 pages

50MCQ_Lecture1[1]

The document consists of a series of multiple-choice questions (MCQs) focused on word embeddings and their applications in natural language processing (NLP). Key topics include the goals of word embeddings, the advantages over one-hot vectors, various embedding methods, and the roles of different components in the Continuous Bag of Words (CBOW) model. It also addresses challenges in training and using word embeddings for tasks like sentiment analysis and machine translation.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 18

50MCQ Lecture1: Embedding Word

1. What is the main goal of word embeddings?


a) To convert text into integers
b) To represent words in a high-dimensional space
c) To capture semantic meaning of words in a low-dimensional space
d) To remove stop words from text

Answer: c) To capture semantic meaning of words in a low-dimensional space

2. Which of the following is a disadvantage of one-hot vectors?


a) They are simple to implement
b) They do not capture semantic relationships between words
c) They are computationally efficient
d) They are suitable for large vocabularies

Answer: b) They do not capture semantic relationships between words

3. What does "semantic distance" mean in word embeddings?


a) The physical distance between words in a document
b) The similarity in meaning between words
c) The frequency of words in a corpus
d) The length of words in characters

Answer: b) The similarity in meaning between words


4. Which of the following is NOT a basic word embedding method?
a) word2vec
b) GloVe
c) BERT
d) fastText

Answer: c) BERT (BERT is an advanced, contextual embedding method)

5. What is the main goal of the CBOW model?


a) To predict the next word in a sentence
b) To predict the center word given its context words
c) To classify text into categories
d) To generate random word vectors

Answer: b) To predict the center word given its context words

6. What is the role of the context window in CBOW?


a) It defines the number of words to be predicted
b) It specifies the number of words surrounding the center word
c) It determines the size of the vocabulary
d) It limits the number of training examples

Answer: b) It specifies the number of words surrounding the center word


7. Which of the following is a key hyperparameter in CBOW?
a) Learning rate
b) Number of layers in the neural network
c) Size of the word embeddings
d) All of the above

Answer: d) All of the above

8. What is the purpose of the ReLU activation function in CBOW?


a) To normalize the input data
b) To introduce non-linearity into the model
c) To reduce the dimensionality of the word embeddings
d) To calculate the loss function

Answer: b) To introduce non-linearity into the model

9. What does the softmax function do in CBOW?


a) It converts the output into probabilities
b) It calculates the loss function
c) It normalizes the input vectors
d) It reduces the dimensionality of the word embeddings

Answer: a) It converts the output into probabilities

10. Which loss function is commonly used in CBOW?


a) Mean squared error
b) Cross-entropy loss
c) Hinge loss
d) Logistic loss

Answer: b) Cross-entropy loss

11. What is the main advantage of word embeddings over one-hot vectors?
a) They are easier to compute
b) They capture semantic relationships between words
c) They require less memory
d) They are faster to train

Answer: b) They capture semantic relationships between words

12. Which of the following is an example of a semantic analogy?


a) "Paris" is to "France" as "Rome" is to "Italy"
b) "Seen" is to "saw" as "been" is to "was"
c) "Wolf" is to "pack" as "bee" is to "swarm"
d) All of the above

Answer: d) All of the above

13. What is the purpose of intrinsic evaluation in word embeddings?


a) To test the embeddings on external tasks
b) To evaluate the embeddings based on their performance in analogies and
clustering
c) To measure the computational efficiency of the embeddings
d) To compare different embedding methods

Answer: b) To evaluate the embeddings based on their performance in analogies


and clustering

14. Which of the following is an example of extrinsic evaluation?


a) Testing word embeddings on named entity recognition
b) Evaluating word embeddings using semantic analogies
c) Visualizing word embeddings in 2D space
d) Clustering words based on their embeddings

Answer: a) Testing word embeddings on named entity recognition

15. What is the main challenge of extrinsic evaluation?


a) It is computationally expensive
b) It is difficult to interpret the results
c) It requires a large amount of labeled data
d) All of the above

Answer: d) All of the above

16. What is the purpose of tokenization in NLP?


a) To convert text into integers
b) To split text into individual words or tokens
c) To remove punctuation from text
d) To normalize text by converting it to lowercase

Answer: b) To split text into individual words or tokens

17. Which of the following is a common preprocessing step in NLP?


a) Converting text to lowercase
b) Removing special characters
c) Handling numbers and punctuation
d) All of the above

Answer: d) All of the above

18. What is the main advantage of using pre-trained word embeddings?


a) They are faster to train
b) They require less data
c) They capture general language patterns
d) All of the above

Answer: d) All of the above

19. Which of the following is a popular pre-trained word embedding model?


a) word2vec
b) GloVe
c) fastText
d) All of the above

Answer: d) All of the above

20. What is the primary difference between word2vec and GloVe?


a) word2vec uses a neural network, while GloVe uses matrix factorization
b) word2vec is faster to train than GloVe
c) GloVe captures global word co-occurrence statistics, while word2vec
captures local context
d) Both a) and c)

Answer: d) Both a) and c)

21. What is the key advantage of fastText over word2vec?


a) It supports out-of-vocabulary words
b) It is faster to train
c) It captures semantic relationships better
d) It requires less memory

Answer: a) It supports out-of-vocabulary words

22. What is the purpose of the sliding window technique in CBOW?


a) To generate training examples by selecting context words around a center
word
b) To reduce the dimensionality of the word embeddings
c) To calculate the loss function
d) To normalize the input data

Answer: a) To generate training examples by selecting context words around a


center word

23. What is the role of the projection layer in CBOW?


a) To convert context words into a single vector
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To convert context words into a single vector

24. What is the main purpose of the hidden layer in CBOW?


a) To capture non-linear relationships between words
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To capture non-linear relationships between words


25. What is the primary purpose of the output layer in CBOW?
a) To predict the center word
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To predict the center word

26. What is the main advantage of using word embeddings in sentiment


analysis?
a) They capture the sentiment of words
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above

Answer: d) All of the above

27. Which of the following is a key challenge in training word embeddings?


a) Choosing the right hyperparameters
b) Handling out-of-vocabulary words
c) Capturing rare words
d) All of the above

Answer: d) All of the above


28. What is the main purpose of the softmax function in CBOW?
a) To convert the output into probabilities
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To convert the output into probabilities

29. What is the main purpose of the cross-entropy loss function in CBOW?
a) To measure the difference between predicted and actual center words
b) To normalize the input data
c) To reduce the dimensionality of the word embeddings
d) To calculate the accuracy of the model

Answer: a) To measure the difference between predicted and actual center


words

30. What is the main advantage of using word embeddings in machine


translation?
a) They capture the meaning of words in different languages
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above
Answer:d) All of the above

31. Which of the following is a key challenge in using word embeddings for
machine translation?
a) Handling out-of-vocabulary words
b) Capturing rare words
c) Aligning words in different languages
d) All of the above

Answer: d) All of the above

32. What is the main purpose of the ReLU activation function in CBOW?
a) To introduce non-linearity into the model
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To introduce non-linearity into the model

33. What is the main purpose of the hidden layer in CBOW?


a) To capture non-linear relationships between words
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To capture non-linear relationships between words

34. What is the main purpose of the output layer in CBOW?


a) To predict the center word
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To predict the center word

35. What is the main advantage of using word embeddings in sentiment


analysis?
a) They capture the sentiment of words
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above

Answer: d) All of the above


36. Which of the following is a key challenge in training word embeddings?
a) Choosing the right hyperparameters
b) Handling out-of-vocabulary words
c) Capturing rare words
d) All of the above

Answer: d) All of the above

37. What is the main purpose of the softmax function in CBOW?


a) To convert the output into probabilities
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To convert the output into probabilities

38. What is the main purpose of the cross-entropy loss function in CBOW?
a) To measure the difference between predicted and actual center words
b) To normalize the input data
c) To reduce the dimensionality of the word embeddings
d) To calculate the accuracy of the model
Answer: a) To measure the difference between predicted and actual center
words

39. What is the main advantage of using word embeddings in machine


translation?
a) They capture the meaning of words in different languages
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above

Answer: d) All of the above

40. Which of the following is a key challenge in using word embeddings for
machine translation?
a) Handling out-of-vocabulary words
b) Capturing rare words
c) Aligning words in different languages
d) All of the above

Answer: d) All of the above

41. What is the main purpose of the ReLU activation function in CBOW?
a) To introduce non-linearity into the model
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To introduce non-linearity into the model


42. What is the main purpose of the hidden layer in CBOW?
a) To capture non-linear relationships between words
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To capture non-linear relationships between words

43. What is the main purpose of the output layer in CBOW?


a) To predict the center word
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To predict the center word

44. What is the main advantage of using word embeddings in sentiment


analysis?
a) They capture the sentiment of words
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above
Answer: d) All of the above

45. Which of the following is a key challenge in training word embeddings?


a) Choosing the right hyperparameters
b) Handling out-of-vocabulary words
c) Capturing rare words
d) All of the above

Answer: d) All of the above

46. What is the main purpose of the softmax function in CBOW?


a) To convert the output into probabilities
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To convert the output into probabilities

47. What is the main purpose of the cross-entropy loss function in CBOW?
a) To measure the difference between predicted and actual center words
b) To normalize the input data
c) To reduce the dimensionality of the word embeddings
d) To calculate the accuracy of the model

Answer: a) To measure the difference between predicted and actual center


words

48. What is the main advantage of using word embeddings in machine


translation?
a) They capture the meaning of words in different languages
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above

Answer: d) All of the above

49. Which of the following is a key challenge in using word embeddings for
machine translation?
a) Handling out-of-vocabulary words
b) Capturing rare words
c) Aligning words in different languages
d) All of the above

Answer: d) All of the above


50. What is the main purpose of the ReLU activation function in CBOW?
a) To introduce non-linearity into the model
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings

Answer: a) To introduce non-linearity into the model

You might also like