50MCQ Lecture1: Embedding Word
1. What is the main goal of word embeddings?
a) To convert text into integers
b) To represent words in a high-dimensional space
c) To capture semantic meaning of words in a low-dimensional space
d) To remove stop words from text
Answer: c) To capture semantic meaning of words in a low-dimensional space
2. Which of the following is a disadvantage of one-hot vectors?
a) They are simple to implement
b) They do not capture semantic relationships between words
c) They are computationally efficient
d) They are suitable for large vocabularies
Answer: b) They do not capture semantic relationships between words
3. What does "semantic distance" mean in word embeddings?
a) The physical distance between words in a document
b) The similarity in meaning between words
c) The frequency of words in a corpus
d) The length of words in characters
Answer: b) The similarity in meaning between words
4. Which of the following is NOT a basic word embedding method?
a) word2vec
b) GloVe
c) BERT
d) fastText
Answer: c) BERT (BERT is an advanced, contextual embedding method)
5. What is the main goal of the CBOW model?
a) To predict the next word in a sentence
b) To predict the center word given its context words
c) To classify text into categories
d) To generate random word vectors
Answer: b) To predict the center word given its context words
6. What is the role of the context window in CBOW?
a) It defines the number of words to be predicted
b) It specifies the number of words surrounding the center word
c) It determines the size of the vocabulary
d) It limits the number of training examples
Answer: b) It specifies the number of words surrounding the center word
7. Which of the following is a key hyperparameter in CBOW?
a) Learning rate
b) Number of layers in the neural network
c) Size of the word embeddings
d) All of the above
Answer: d) All of the above
8. What is the purpose of the ReLU activation function in CBOW?
a) To normalize the input data
b) To introduce non-linearity into the model
c) To reduce the dimensionality of the word embeddings
d) To calculate the loss function
Answer: b) To introduce non-linearity into the model
9. What does the softmax function do in CBOW?
a) It converts the output into probabilities
b) It calculates the loss function
c) It normalizes the input vectors
d) It reduces the dimensionality of the word embeddings
Answer: a) It converts the output into probabilities
10. Which loss function is commonly used in CBOW?
a) Mean squared error
b) Cross-entropy loss
c) Hinge loss
d) Logistic loss
Answer: b) Cross-entropy loss
11. What is the main advantage of word embeddings over one-hot vectors?
a) They are easier to compute
b) They capture semantic relationships between words
c) They require less memory
d) They are faster to train
Answer: b) They capture semantic relationships between words
12. Which of the following is an example of a semantic analogy?
a) "Paris" is to "France" as "Rome" is to "Italy"
b) "Seen" is to "saw" as "been" is to "was"
c) "Wolf" is to "pack" as "bee" is to "swarm"
d) All of the above
Answer: d) All of the above
13. What is the purpose of intrinsic evaluation in word embeddings?
a) To test the embeddings on external tasks
b) To evaluate the embeddings based on their performance in analogies and
clustering
c) To measure the computational efficiency of the embeddings
d) To compare different embedding methods
Answer: b) To evaluate the embeddings based on their performance in analogies
and clustering
14. Which of the following is an example of extrinsic evaluation?
a) Testing word embeddings on named entity recognition
b) Evaluating word embeddings using semantic analogies
c) Visualizing word embeddings in 2D space
d) Clustering words based on their embeddings
Answer: a) Testing word embeddings on named entity recognition
15. What is the main challenge of extrinsic evaluation?
a) It is computationally expensive
b) It is difficult to interpret the results
c) It requires a large amount of labeled data
d) All of the above
Answer: d) All of the above
16. What is the purpose of tokenization in NLP?
a) To convert text into integers
b) To split text into individual words or tokens
c) To remove punctuation from text
d) To normalize text by converting it to lowercase
Answer: b) To split text into individual words or tokens
17. Which of the following is a common preprocessing step in NLP?
a) Converting text to lowercase
b) Removing special characters
c) Handling numbers and punctuation
d) All of the above
Answer: d) All of the above
18. What is the main advantage of using pre-trained word embeddings?
a) They are faster to train
b) They require less data
c) They capture general language patterns
d) All of the above
Answer: d) All of the above
19. Which of the following is a popular pre-trained word embedding model?
a) word2vec
b) GloVe
c) fastText
d) All of the above
Answer: d) All of the above
20. What is the primary difference between word2vec and GloVe?
a) word2vec uses a neural network, while GloVe uses matrix factorization
b) word2vec is faster to train than GloVe
c) GloVe captures global word co-occurrence statistics, while word2vec
captures local context
d) Both a) and c)
Answer: d) Both a) and c)
21. What is the key advantage of fastText over word2vec?
a) It supports out-of-vocabulary words
b) It is faster to train
c) It captures semantic relationships better
d) It requires less memory
Answer: a) It supports out-of-vocabulary words
22. What is the purpose of the sliding window technique in CBOW?
a) To generate training examples by selecting context words around a center
word
b) To reduce the dimensionality of the word embeddings
c) To calculate the loss function
d) To normalize the input data
Answer: a) To generate training examples by selecting context words around a
center word
23. What is the role of the projection layer in CBOW?
a) To convert context words into a single vector
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To convert context words into a single vector
24. What is the main purpose of the hidden layer in CBOW?
a) To capture non-linear relationships between words
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To capture non-linear relationships between words
25. What is the primary purpose of the output layer in CBOW?
a) To predict the center word
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To predict the center word
26. What is the main advantage of using word embeddings in sentiment
analysis?
a) They capture the sentiment of words
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above
Answer: d) All of the above
27. Which of the following is a key challenge in training word embeddings?
a) Choosing the right hyperparameters
b) Handling out-of-vocabulary words
c) Capturing rare words
d) All of the above
Answer: d) All of the above
28. What is the main purpose of the softmax function in CBOW?
a) To convert the output into probabilities
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To convert the output into probabilities
29. What is the main purpose of the cross-entropy loss function in CBOW?
a) To measure the difference between predicted and actual center words
b) To normalize the input data
c) To reduce the dimensionality of the word embeddings
d) To calculate the accuracy of the model
Answer: a) To measure the difference between predicted and actual center
words
30. What is the main advantage of using word embeddings in machine
translation?
a) They capture the meaning of words in different languages
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above
Answer:d) All of the above
31. Which of the following is a key challenge in using word embeddings for
machine translation?
a) Handling out-of-vocabulary words
b) Capturing rare words
c) Aligning words in different languages
d) All of the above
Answer: d) All of the above
32. What is the main purpose of the ReLU activation function in CBOW?
a) To introduce non-linearity into the model
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To introduce non-linearity into the model
33. What is the main purpose of the hidden layer in CBOW?
a) To capture non-linear relationships between words
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To capture non-linear relationships between words
34. What is the main purpose of the output layer in CBOW?
a) To predict the center word
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To predict the center word
35. What is the main advantage of using word embeddings in sentiment
analysis?
a) They capture the sentiment of words
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above
Answer: d) All of the above
36. Which of the following is a key challenge in training word embeddings?
a) Choosing the right hyperparameters
b) Handling out-of-vocabulary words
c) Capturing rare words
d) All of the above
Answer: d) All of the above
37. What is the main purpose of the softmax function in CBOW?
a) To convert the output into probabilities
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To convert the output into probabilities
38. What is the main purpose of the cross-entropy loss function in CBOW?
a) To measure the difference between predicted and actual center words
b) To normalize the input data
c) To reduce the dimensionality of the word embeddings
d) To calculate the accuracy of the model
Answer: a) To measure the difference between predicted and actual center
words
39. What is the main advantage of using word embeddings in machine
translation?
a) They capture the meaning of words in different languages
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above
Answer: d) All of the above
40. Which of the following is a key challenge in using word embeddings for
machine translation?
a) Handling out-of-vocabulary words
b) Capturing rare words
c) Aligning words in different languages
d) All of the above
Answer: d) All of the above
41. What is the main purpose of the ReLU activation function in CBOW?
a) To introduce non-linearity into the model
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To introduce non-linearity into the model
42. What is the main purpose of the hidden layer in CBOW?
a) To capture non-linear relationships between words
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To capture non-linear relationships between words
43. What is the main purpose of the output layer in CBOW?
a) To predict the center word
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To predict the center word
44. What is the main advantage of using word embeddings in sentiment
analysis?
a) They capture the sentiment of words
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above
Answer: d) All of the above
45. Which of the following is a key challenge in training word embeddings?
a) Choosing the right hyperparameters
b) Handling out-of-vocabulary words
c) Capturing rare words
d) All of the above
Answer: d) All of the above
46. What is the main purpose of the softmax function in CBOW?
a) To convert the output into probabilities
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To convert the output into probabilities
47. What is the main purpose of the cross-entropy loss function in CBOW?
a) To measure the difference between predicted and actual center words
b) To normalize the input data
c) To reduce the dimensionality of the word embeddings
d) To calculate the accuracy of the model
Answer: a) To measure the difference between predicted and actual center
words
48. What is the main advantage of using word embeddings in machine
translation?
a) They capture the meaning of words in different languages
b) They reduce the dimensionality of the input data
c) They improve the accuracy of the model
d) All of the above
Answer: d) All of the above
49. Which of the following is a key challenge in using word embeddings for
machine translation?
a) Handling out-of-vocabulary words
b) Capturing rare words
c) Aligning words in different languages
d) All of the above
Answer: d) All of the above
50. What is the main purpose of the ReLU activation function in CBOW?
a) To introduce non-linearity into the model
b) To calculate the loss function
c) To normalize the input data
d) To reduce the dimensionality of the word embeddings
Answer: a) To introduce non-linearity into the model