5 items with this tag.
papers
Paper note on shrinking and retargeting the tokenizer and embedding table to a domain so the model uses fewer vocabulary bytes and shorter sequences.
frontiers
Frontier synthesis on why tokenizer research for compact models should be treated as a joint vocabulary, logits, and artifact-budget problem rather than a token-count problem.
lanes
Tokenization is part of the budget story, not just a preprocessing detail.
papers
Paper note on replacing a pretrained model tokenizer while retraining only embeddings and the LM head.
papers
Paper note on reducing output-layer memory and logits cost by restructuring vocabulary prediction.