GGUF
llama.cpp

[THINK] token not recognized in llama-cpp-python

#1
by Kenshiro-28 - opened

It looks the [THINK] token is not recognized, as this line:

print(len(model.tokenize(b"[THINK]", add_bos=False)));

Prints a 4 instead of 1

I noticed that some regions near the top of the tensor slice look "flat" several tensor slices have nearly identical values
https://www.reddit.com/r/MistralAI/comments/1o0kvpe/weird_repeating_values_in_magistralsmall2509_gguf/

Kenshiro-28 changed discussion status to closed

Sign up or log in to comment