gguf when? cmon, its been 11 min already!

#2
by Hanswalter - opened

x

lol well darn, i had plans today... oof... as a quantizer, i wonder if i should wait for the -Instruct ? is that out yet? lol...

Better call @bartowski

@MarinaraSpaghetti

I'll put up the Bat-towski signal!

@ubergarm I was hoping to see you in one of these threads :D

+1 gguf please

wait for instruct model, not sure how gguf of the base model could be usefull for personal usage

Base models are good for creative writing.

This comment has been hidden (marked as Abuse)

lol well darn, i had plans today... oof... as a quantizer, i wonder if i should wait for the -Instruct ? is that out yet? lol...

How dare you have plans when ds puts out a new model!!! 😂

"Why is the GGUF so late it's been 20 seconds already!"

i think lets wait for instruct version. I am very patient. very very very patient.

I think llama.cpp needs to be updated first.

I figured out how to create the bf16 safetensors, now I'm creating the bf16 gguf. We'll see.

Yeah, seems like it needs some changes to llama.cpp. I got it inferring but the chat template seems messed up.

I'm throwing a Q4_K_M up soon while I work on imatrix and further quants

@createthis it's also a base model so chatting is not going to be as reliable without giving it a multi turn prompt

@bartowski Thanks for the llama-cli example. TIL.

@

@MarinaraSpaghetti

I'll put up the Bat-towski signal!

when will be in ik_llama :p

Team mradermacher is now generating quants. You can follow the progress on the status page under https://hf.tst.eu/status.html. The first static quants should appear under https://huggingface.co/mradermacher/DeepSeek-V3.1-Base-GGUF within the next few hours.

when will be in ik_llama :p

Yeah, I need ik_llama to fit decent quality Deepseek on my hardware too.

I'm making an iq2_ks for myself ( @ubergarm 's cookbook and his calibration dataset for the imatrix). I'll upload it if nobody else has made anything better by the time it's done.

I've also uploaded the upcast bf16 ggufs: gghfez/DeepSeek-V3.1-Base-256x21B-BF16 if it helps anyone else making ik_llama quants.

@gghfez

Thanks for providing something for -Base for people to try out, keep in mind that imatrix was made for an earlier version so might not be so accurate using against a different version of the weights.

I'm gonna start working on the Instruct now that it is ready: https://huggingface.co/deepseek-ai/DeepSeek-V3.1

Sign up or log in to comment