Mistral-small
#19 opened 6 months ago
by
Melkiss
Add chat template to tokenizer config
#18 opened 6 months ago
by
mrfakename

Mistral3ForConditionalGeneration has no vLLM implementation and the Transformers implementation is not compatible with vLLM. Try setting VLLM_USE_V1=0.
👍
4
3
#16 opened 6 months ago
by
pedrojfb99
set model_max_length to the maximum length of model context (131072 tokens)
#15 opened 6 months ago
by
x0wllaar

Problem with `mistral3` when loading the model
7
#14 opened 6 months ago
by
r3lativo
Add chat_template to tokenizer_config.json
👍
1
1
#11 opened 6 months ago
by
bethrezen

Can't wait for HF? try chatllm.cpp
🚀
👍
2
6
#7 opened 6 months ago
by
J22
You did it again...
🚀
39
#4 opened 6 months ago
by
MrDevolver

HF Format?
🧠
❤️
33
41
#2 opened 6 months ago
by
bartowski
