fastText embeddings exploit subword information to construct word embeddings. Representations are learnt of character $n$-grams, and words represented as the sum of the $n$-gram vectors. This extends the word2vec type models with subword information. This helps the embeddings understand suffixes and prefixes. Once a word is represented using character $n$-grams, a skipgram model is trained to learn the embeddings.
Source: Enriching Word Vectors with Subword InformationPaper | Code | Results | Date | Stars |
---|
Task | Papers | Share |
---|---|---|
Text Classification | 33 | 7.89% |
General Classification | 28 | 6.70% |
Sentence | 24 | 5.74% |
Sentiment Analysis | 22 | 5.26% |
Classification | 16 | 3.83% |
Named Entity Recognition (NER) | 15 | 3.59% |
Language Modelling | 12 | 2.87% |
Word Similarity | 11 | 2.63% |
Clustering | 7 | 1.67% |
Component | Type |
|
---|---|---|
🤖 No Components Found | You can add them if they exist; e.g. Mask R-CNN uses RoIAlign |