
Tokenomics of Large Language Models - Gradient Street #003
A practical tour of tokenisation for LLMs. Why word-level fails, how subwords and BPE work, and what modern tokenisers look like, with tips and resources. Plus a brief personal update.
A practical tour of tokenisation for LLMs. Why word-level fails, how subwords and BPE work, and what modern tokenisers look like, with tips and resources. Plus a brief personal update.
An accessible tour of word embeddings, from Word2Vec to GloVe, with intuition, a touch of maths, and practical notes on training and using them. Includes caveats, examples, and where modern models fit.
Do machines understand words? And if not exactly, do they use Google Translate?
What can I do to learn better? What ways can I expose myself? And are we really lucky that we are born in Europe in modern times?
What can I do to learn better? What ways can I expose myself? And are we really lucky that we are born in Europe in modern times?
How being a data nerd helps me manage my weight and hit my goals. Progress over perfection.