ELMo
An editor has performed a search and found that sufficient sources exist to establish the subject's notability. These sources can be used to expand the article and may be described in edit summaries or found on the talk page. The article may include original research, or omit significant information about the subject. (September 2020) |
ELMo ("Embeddings from Language Model") is a word embedding method for representing a sequence of words as a corresponding sequence of vectors.[1] Character-level tokens are taken as the inputs to a bi-directional LSTM which produces word-level embeddings. Like BERT (but unlike the word embeddings produced by "Bag of Words" approaches, and earlier vector approaches such as Word2Vec and GloVe), ELMo embeddings are context-sensitive, producing different representations for words that share the same spelling but have different meanings (homonyms) such as "bank" in "river bank" and "bank balance".
It was created by researchers at the Allen Institute for Artificial Intelligence[2] and University of Washington.
References[edit]
- ^ Peters ME, Neumann M, Iyyer M, Gardner M, Clark C, Lee K, Zettlemoyer L (2018). "Deep contextualized word representations". arXiv:1802.05365 [cs.CL].
- ^ "AllenNLP - ELMo — Allen Institute for AI".