Learn Before
Concept
Word Embeddings
Different from SRE models, NRE mainly utilizes word embeddings and position embeddings instead of hand-craft features as inputs. Word embeddings are the most used input representations in NLP, which encode the semantic meaning of words into vectors.
0
1
Updated 2022-06-05
Tags
Data Science