Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
59.7
TFLOPS
26
1
9
s
Tom-Neverwinter
Follow
LeroyDyer's profile picture
21world's profile picture
2 followers
·
16 following
Tom-Neverwinter
AI & ML interests
Making improvements to help the world.
Recent Activity
new
activity
4 days ago
perplexity-ai/r1-1776:
Was this Model Needed?
new
activity
10 days ago
NousResearch/DeepHermes-3-Llama-3-8B-Preview-GGUF:
Engage the hype engines anther hermes has arrived!!!
reacted
to
csabakecskemeti
's
post
with 🔥
about 2 months ago
I've built a small utility to split safetensors file by file. The issue/need came up when I've tried to convert the new Deepseek V3 model from FP8 to BF16. The only Ada architecture GPU I have is an RTX 4080 and the 16GB vram was just wasn't enough for the conversion. BTW: I'll upload the bf16 version here: https://huggingface.co/DevQuasar/deepseek-ai.DeepSeek-V3-Base-bf16 (it will take a while - days with my upload speed) If anyone has access the resources to test it I'd appreciate a feedback if it's working or not. The tool, is available from here: https://github.com/csabakecskemeti/ai_utils/blob/main/safetensor_splitter.py It's splitting every file to n pieces by the layers if possible, and create a new "model.safetensors.index.json" file. I've tested it with Llama 3.1 8B and multiple split sizes, and validated by using inference pipeline. use `--help` for usage Please note current version expects the model is already multiple file and have a "model.safetensors.index.json" layer-safetensor mapping file.
View all activity
Organizations
None yet
models
4
Sort: Recently updated
Tom-Neverwinter/ew-lora
Updated
Aug 16, 2024
•
4
Tom-Neverwinter/ts-lora
Updated
Aug 16, 2024
•
4
Tom-Neverwinter/cr-lora
Updated
Aug 16, 2024
•
2
Tom-Neverwinter/sw-lora
Updated
Aug 16, 2024
•
3
datasets
None public yet