From Ether to Syntax: A Meta-Analytic Exploration of Linguistic Algorithmic Landscapes

#6
by mradermacher - opened

continued....

mradermacher changed discussion status to closed

Here a compleate list of the newly added architectures.

The non-mm-archs are picked up automatically when llama is updated (rather, nothing checks for these archs, other than the script that shows me daily models).

Nice. Will do in caser you forgot any vision/audio architecture.

In case yopu need it, the list/regexc is currently in /llmjob/share/llmjob.pm - search for is_vision

Also, vision is mradermacher code for multi-modal from now on.

Bert based architectures seem to be incredible

I might exclude them from the daily list for that reason, and them being likely not popular with the people who consume ggufs. (and most fail because small models tend to have custom tokenizers).

Nice I just discover an easy way to requeue previously failed archidectures:

Yup, shell-greppable logs for the win.

Update: oh, it's not even the real log file, "just" the llmc why transform of it.

@RichardErkhov vision models should not be queued to rich1 unless they arte not being detected as such (and then no vision extraction should happen).

The non-vision jobs are limited to 32GB ram, too. No clue what happened. Very troubling.

However, this morning, only besteffort models were queued on rich1. Who knows what nico queued...

well, good to know. usually you take like 4-8gb, but something went wrong today. Peak recorded by proxmox was 24gb (so I assume it was even higher, but due to total OOM, it might not have recorded full number. I added swap on root just in case this happens again so at least other things on server dont die haha

llmc audit besteffort skips the besteffort models for me.

Please restart Audio-Reasoner imatrix computation. I killed it earlier today because it ran on CPU. I'm still not sure what makes GPUs occasionally temporary disappear but seams related to them being used on a different container.

llmc audit besteffort skips the besteffort models for me.

Right, arguments were not passed to llmjob audit. Should be fixed now.

@RichardErkhov

Peak recorded by proxmox was 24gb

Well, given that I was officially allowed to use 64GB, 24GB seems absolutely normal. So what is the new limit? 24GB will only allow one quant, and maybe not even that.

@RichardErkhov wow, I admit I totally forgot. I've "disabled" rich1, and it should be finished (only 3 jobs remaining) Thanks foer the heads-up

@nicoboss I don't think S1-Base-671B and cogito-v2-preview-deepseek-671B-MoE should be besteffort?

@mradermacherhi again, as soon as you are done with rich1 and ready for a backup, please just pause it or even shutdown the container. I will do everything else needed with nico

@nicoboss https://huggingface.co/chutesai/Qwen3-235B-A22B-Instruct-2507-1M what are your thoughts about this model? the only thing we have in it's favour is the "-1M"

@RichardErkhov done, it's shut down

thank you, moving now =)

@nicoboss https://huggingface.co/chutesai/Qwen3-235B-A22B-Instruct-2507-1M what are your thoughts about this model? the only thing we have in it's favour is the "-1M"

It's a hash identical copy of https://huggingface.co/Qwen/Qwen3-235B-A22B-Instruct-2507 so let's not waste any resources on this garbage.

Please mark DeepSeek-V3.1-Base for Q8 RPC imatrix computation so I can start it once static quants are done. That model is supper exciting. It’s quite a massive improvement compared to DeepSeek V3 but also slightly more censored.

Sign up or log in to comment