Skip to content

Latest commit

 

History

History
40 lines (21 loc) · 1022 Bytes

01_Word_Vectors.md

File metadata and controls

40 lines (21 loc) · 1022 Bytes

01 Word Vectors

Word2vec: Overview

Word2vec (Mikolov et al. 2013) is a framework for learning word vectors

Idea:

  • We have a large corpus (“body”) of text
  • Every word in a fixed vocabulary is represented by a vector
  • Go through each position t in the text, which has a center word c and context (“outside”) words o
  • Use the similarity of the word vectors for c and o to calculate the probability of o given c (or vice versa)
  • Keep adjusting the word vectors to maximize this probability

Objective function

Prediction function

Word2vec derivations of gradient

Task:

  • 手推Word2Vec损失函数与梯度下降:

Link