The model does have the format specified but there is no _one_ standard. For this model it’s defined in the [
tokenizer_config.json [0]. As for llama.cpp they seem to be using a more type safe approach to reading the arguments.
[0] https://huggingface.co/google/gemma-4-31B-it/blob/main/token...