eos_token becomes <|im_end|> after GGUF conversion with llama.cpp, and generation never terminates

#1
by mmnga - opened

I’d like to report an issue I encountered when converting a model to GGUF with llama.cpp: the eos_token is set to <|im_end|>, which prevents generation from terminating properly.
Updating tokenizer_config.json as follows resolves the problem:

  "eos_token": "<|im_end|>",

  "eos_token": "<|end▁of▁sentence|>",

After making this change, the model stops correctly at the end-of-sentence token.
Additionally, I hope this information is helpful.

Sign up or log in to comment