Quiz: Applying BERT to Other Languages

Take a short quiz to test your understanding of the multilingual BERT model.

1

What is the primary motivation behind oversampling and undersampling in multilingual BERT?

A)

To prioritize high-resource languages for training.

B)

To exclude the low-resource languages from training.

C)

To maintain a balanced data distribution in each language.

D)

To randomly shuffle the training data to achieve balance.

Question 1 of 80 attempted

Get hands-on with 1200+ tech skills courses.