longform: Global Vectodoc

Usecase of GloVe

It has the same Usecase as Word2Vec but it uses its global context for the embedding of the words into multidimensionalv vectors.

  1. Build the Term cooccurence matrix
  2. Matrix factorization
  3. Vectorrepresenation of Words
  4. Use similarity like cosine sim
  5. Use Probality Ratio P(k|word) and with that it calculate the relation to the words getting the Words closer toghether.
  6. backprop: Optimizes probality ratio

GloVe: Global Vectors for Word Representation

This example is highly simplified. In real scenarios, the matrix is large, sparse, and constructed from extensive text corpora to capture meaningful statistical relationships between words effectively.