Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
π Abstract
Abstract: Language models (LMs) are bound to their tokenizer, which maps raw text to a
sequence of vocabulary items (tokens). This restricts their flexibility: for
example, LMs trained primarily on English may still perform well in other
natural and programming languages, but have vastly decreased efficiency due to
their English-centric tokenizer. To mitigate this, we should be able to swap
the original LM tokenizer with an arbitrary one, on the fly, without degrading
performance. Hence, in this work we define a new problem: Zero-Shot Tokenizer
Transfer (ZeTT). The challenge at the core of ZeTT is finding embeddings for
the tokens in the vocabulary of the new tokenizer. Since prior heuristics for
initializing embeddings often perform at chance level in a ZeTT setting, we
propose a new solution: we train a hypernetwork taking a tokenizer as input and
predicting the corresponding embeddings. We empirically demonstrate that the
hypernetwork generalizes to new tokenizers both with encoder (e.g., XLM-R) and
decoder LLMs (e.g., Mistral-7B). Our method comes close to the original models'
performance in cross-lingual and coding tasks while markedly reducing the
length of the tokenized sequence. We also find that the remaining gap can be
quickly closed by continued training on less than 1B tokens. Finally, we show
that a ZeTT hypernetwork trained for a base (L)LM can also be applied to
fine-tuned variants without extra training. Overall, our results make
substantial strides toward detaching LMs from their tokenizer.
Authors (3)
Benjamin Minixhofer
Edoardo Maria Ponti
Ivan VuliΔ
Key Contributions
This paper defines and addresses the problem of Zero-Shot Tokenizer Transfer (ZeTT), enabling language models to swap their original tokenizer with an arbitrary one without performance degradation. It proposes a novel solution using a hypernetwork that takes a tokenizer as input and predicts the corresponding embeddings. This approach significantly improves efficiency for LMs operating on languages or codebases different from their training data, overcoming the limitations of previous heuristics.
Business Value
Allows for more flexible and efficient deployment of large language models across diverse languages and programming languages without costly retraining, significantly reducing operational costs and expanding the applicability of LLMs.