Here are all the words used in the notes, and their meaning according to slides.
- Corpora: courpus, a computer readable collection
- Utterance: uhm, ehm…
- Word Types: set of distinct words in a corpus , where (e.g. “I like fish and I like chips”: word types = 5, because we must not count the duplicates)
- Tokens: number of word in a phrase (e.g. “I like fish and I like chips”: tokens = 7)
- Lemmatization: extract root from a word
- Minimization: see Loss Function
- Loss Function: how good a model is. The lower the better, it means that is very accurate in predicting
- Softmax: soft probability smaller than x, max amplifies probabilities of target . Maps arbitrary values to probability distribution
- Gradient: how much vectors need to be in order to reduce errors