Tokenization is the first step toward transforming text into machine-friendly units. Karpathy touches on widely used ...