Difference between Doc2Vec and BERT
I am trying to understand the difference between Doc2Vec and BERT. I do understand that doc2vec uses a paragraph ID which also serves as a paragraph vector. I am not sure though if that paragraph ID serves in better able to understand the context in that vector?
Moreover, BERT definitely understands the context and attributes different vectors for words such as Bank. for instance,
- I robbed a bank
- I was sitting by the bank of a river.
BERT would allocate different vectors for the word BANK here. Trying to understand if doc2vec also gets this context since the paragraph id would be different here (for doc2vec). Can anyone please help with this?
Topic doc2vec bert transformer nlp machine-learning
Category Data Science