When using vLLM, a missing chat_template.json results in repeated fetches of the chat template (presumably from tokenizer_config.json), for every single row that the chat template needs to be applied to.

Ready to merge
This branch is ready to get merged automatically.

Sign up or log in to comment