Quiz

1

Why do we use word embeddings?

A)

They’re used to represent a sequence of words as a single continuous vector

B)

Word embeddings are easier to look up than regular tokenized IDs

C)

It makes training models for NLP tasks quicker and more efficient

D)

As a more meaningful way to capture the connections between vocabulary words

Question 1 of 40 attempted

Get hands-on with 1400+ tech skills courses.