Jorge Alonso PRO

oieieio
Β·

AI & ML interests

AI/ML

Recent Activity

updated a Space about 19 hours ago
oieieio/meta-llama-Llama-3.2-1B-Instruct
updated a Space about 20 hours ago
oieieio/First_agent_template
published a Space 6 days ago
oieieio/meta-llama-Llama-3.2-1B-Instruct
View all activity

Organizations

Hugging Face Discord Community's profile picture OIEIEIO's profile picture Lambda Go Labs's profile picture

oieieio's activity

replied to fdaudens's post 10 days ago
view reply

If you're building with AI at any scale, definitely worth checking out. Yes! Looks great!

reacted to fdaudens's post with πŸ”₯ 10 days ago
view post
Post
2659
⭐️ The AI Energy Score project just launched - this is a game-changer for making informed decisions about AI deployment.

You can now see exactly how much energy your chosen model will consume, with a simple 5-star rating system. Think appliance energy labels, but for AI.

Looking at transcription models on the leaderboard is fascinating: choosing between whisper-tiny or whisper-large-v3 can make a 7x difference. Real-time data on these tradeoffs changes everything.

166 models already evaluated across 10 different tasks, from text generation to image classification. The whole thing is public and you can submit your own models to test.

Why this matters:
- Teams can pick efficient models that still get the job done
- Developers can optimize for energy use from day one
- Organizations can finally predict their AI environmental impact

If you're building with AI at any scale, definitely worth checking out.

πŸ‘‰ leaderboard: https://lnkd.in/esrSxetj
πŸ‘‰ blog post: https://lnkd.in/eFJvzHi8

Huge work led by @sasha with @bgamazay @yjernite @sarahooker @regisss @meg
  • 1 reply
Β·
reacted to ZennyKenny's post with πŸ€— 11 days ago
view post
Post
3414
I've completed the first unit of the just-launched Hugging Face Agents Course. I would highly recommend it, even for experienced builders, because it is a great walkthrough of the smolagents library and toolkit.
reacted to davidberenstein1957's post with πŸ€— 11 days ago
view post
Post
3106
πŸš€ Find banger tools for your smolagents!

I created the Tools gallery, which makes tools specifically developed by/for smolagents searchable and visible. This will help with:
- inspiration
- best practices
- finding cool tools

Space: davidberenstein1957/smolagents-and-tools
  • 1 reply
Β·
upvoted an article 13 days ago
reacted to lewtun's post with ❀️πŸ”₯ 13 days ago
view post
Post
4475
Introducing OpenR1-Math-220k!

open-r1/OpenR1-Math-220k

The community has been busy distilling DeepSeek-R1 from inference providers, but we decided to have a go at doing it ourselves from scratch πŸ’ͺ

What’s new compared to existing reasoning datasets?

β™Ύ Based on AI-MO/NuminaMath-1.5: we focus on math reasoning traces and generate answers for problems in NuminaMath 1.5, an improved version of the popular NuminaMath-CoT dataset.

🐳 800k R1 reasoning traces: We generate two answers for 400k problems using DeepSeek R1. The filtered dataset contains 220k problems with correct reasoning traces.

πŸ“€ 512 H100s running locally: Instead of relying on an API, we leverage vLLM and SGLang to run generations locally on our science cluster, generating 180k reasoning traces per day.

⏳ Automated filtering: We apply Math Verify to only retain problems with at least one correct answer. We also leverage Llama3.3-70B-Instruct as a judge to retrieve more correct examples (e.g for cases with malformed answers that can’t be verified with a rules-based parser)

πŸ“Š We match the performance of DeepSeek-Distill-Qwen-7B by finetuning Qwen-7B-Math-Instruct on our dataset.

πŸ”Ž Read our blog post for all the nitty gritty details: https://huggingface.co/blog/open-r1/update-2
reacted to mkurman's post with πŸ‘ 15 days ago
view post
Post
1577
Blurred-Thoughts Supervised-Finetuning πŸ™ˆ

After hours of working with GitHub Copilot to organize the code, I'm keen to announce the release of Blurred Thoughts Supervised-Finetuning (BT-SFT), a new method for fine-tuning LLMs to produce more diverse and creative responses.

BT-SFT introduces:
βœ… Smart tokenization method randomly masks tokens within <think> ... </think> tags, promoting the model to generate diverse responses that align better with its probability distribution instead of memorizing the thought process from distilled data.
βœ… Reward function that ensures responses are well-structured.

Explore and contribute to the project available in my GitHub repository:
https://github.com/mkurman/blurred-thoughts-SFT

Keep me updated on your experiments with BT-SFT! 🐐
upvoted an article 17 days ago
view article
Article

Train 400x faster Static Embedding Models with Sentence Transformers

β€’ 148
New activity in IamCreateAI/Ruyi-Mini-7B about 2 months ago