GO:OD:AM's picture
3 21

GO:OD:AM PRO

tegridydev

AI & ML interests

Mechanistic Interpretability (MI) Research & sp00ky code stuff

Recent Activity

Organizations

None yet

tegridydev's activity

reacted to their post with πŸ€—πŸ”₯ 1 day ago
view post
Post
1926
Open Source AI Agents | Github/Repo List | [2025]

https://huggingface.co/blog/tegridydev/open-source-ai-agents-directory

Check out the article & Follow, bookmark, save the tab as I will be updating it <3
(using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
reacted to their post with ❀️ 2 days ago
view post
Post
1926
Open Source AI Agents | Github/Repo List | [2025]

https://huggingface.co/blog/tegridydev/open-source-ai-agents-directory

Check out the article & Follow, bookmark, save the tab as I will be updating it <3
(using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
posted an update 2 days ago
view post
Post
1926
Open Source AI Agents | Github/Repo List | [2025]

https://huggingface.co/blog/tegridydev/open-source-ai-agents-directory

Check out the article & Follow, bookmark, save the tab as I will be updating it <3
(using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
upvoted an article 2 days ago
published an article 2 days ago
reacted to their post with ❀️ 6 days ago
view post
Post
1892
WTF is Fine-Tuning? (intro4devs)

Fine-tuning your LLM is like min-maxing your ARPG hero so you can push high-level dungeons and get the most out of your build/gear... Makes sense, right? πŸ˜ƒ

Here's a cheat sheet for devs (but open to anyone!)

---

TL;DR

- Full Fine-Tuning: Max performance, high resource needs, best reliability.
- PEFT: Efficient, cost-effective, mainstream, enhanced by AutoML.
- Instruction Fine-Tuning: Ideal for command-following AI, often combined with RLHF and CoT.
- RAFT: Best for fact-grounded models with dynamic retrieval.
- RLHF: Produces ethical, high-quality conversational AI, but expensive.

Choose wisely and match your approach to your task, budget, and deployment constraints.

I just posted the full extended article here
if you want to continue reading >>>

https://huggingface.co/blog/tegridydev/fine-tuning-dev-intro-2025
posted an update 6 days ago
view post
Post
1892
WTF is Fine-Tuning? (intro4devs)

Fine-tuning your LLM is like min-maxing your ARPG hero so you can push high-level dungeons and get the most out of your build/gear... Makes sense, right? πŸ˜ƒ

Here's a cheat sheet for devs (but open to anyone!)

---

TL;DR

- Full Fine-Tuning: Max performance, high resource needs, best reliability.
- PEFT: Efficient, cost-effective, mainstream, enhanced by AutoML.
- Instruction Fine-Tuning: Ideal for command-following AI, often combined with RLHF and CoT.
- RAFT: Best for fact-grounded models with dynamic retrieval.
- RLHF: Produces ethical, high-quality conversational AI, but expensive.

Choose wisely and match your approach to your task, budget, and deployment constraints.

I just posted the full extended article here
if you want to continue reading >>>

https://huggingface.co/blog/tegridydev/fine-tuning-dev-intro-2025
upvoted an article 6 days ago
published an article 6 days ago
upvoted an article 22 days ago
view article
Article

LLM Dataset Formats 101: A No‐BS Guide for Hugging Face Devs

By tegridydev β€’
β€’ 5
published an article 22 days ago
view article
Article

LLM Dataset Formats 101: A No‐BS Guide for Hugging Face Devs

By tegridydev β€’
β€’ 5
reacted to their post with ❀️ 22 days ago
view post
Post
1437
Open-MalSec v0.1 – Open-Source Cybersecurity Dataset

Evening! 🫑

πŸ“‚ Just uploaded an early-stage open-source cybersecurity dataset focused on phishing, scams, and malware-related text samples.

This is the base version (v0.1)β€”a few structured sample files. Full dataset builds will come over the next few weeks.

πŸ”— Dataset link:

tegridydev/open-malsec

πŸ” What’s in v0.1?
A few structured scam examples (text-based)
Covers DeFi, crypto, phishing, and social engineering
Initial labelling format for scam classification

⚠️ This is not a full dataset yet (samples are currently available). Just establishing the structure + getting feedback.

πŸ“‚ Current Schema & Labelling Approach
"instruction" β†’ Task prompt (e.g., "Evaluate this message for scams")
"input" β†’ Source & message details (e.g., Telegram post, Tweet)
"output" β†’ Scam classification & risk indicators

πŸ—‚οΈ Current v0.1 Sample Categories
Crypto Scams β†’ Meme token pump & dumps, fake DeFi projects
Phishing β†’ Suspicious finance/social media messages
Social Engineering β†’ Manipulative messages exploiting trust

πŸ”œ Next Steps
- Expanding datasets with more phishing & malware examples
- Refining schema & annotation quality
- Open to feedback, contributions, and suggestions

If this is something you might find useful, bookmark/follow/like the dataset repo <3

πŸ’¬ Thoughts, feedback, and ideas are always welcome! Drop a comment or DMs are open πŸ€™
posted an update 23 days ago
view post
Post
1437
Open-MalSec v0.1 – Open-Source Cybersecurity Dataset

Evening! 🫑

πŸ“‚ Just uploaded an early-stage open-source cybersecurity dataset focused on phishing, scams, and malware-related text samples.

This is the base version (v0.1)β€”a few structured sample files. Full dataset builds will come over the next few weeks.

πŸ”— Dataset link:

tegridydev/open-malsec

πŸ” What’s in v0.1?
A few structured scam examples (text-based)
Covers DeFi, crypto, phishing, and social engineering
Initial labelling format for scam classification

⚠️ This is not a full dataset yet (samples are currently available). Just establishing the structure + getting feedback.

πŸ“‚ Current Schema & Labelling Approach
"instruction" β†’ Task prompt (e.g., "Evaluate this message for scams")
"input" β†’ Source & message details (e.g., Telegram post, Tweet)
"output" β†’ Scam classification & risk indicators

πŸ—‚οΈ Current v0.1 Sample Categories
Crypto Scams β†’ Meme token pump & dumps, fake DeFi projects
Phishing β†’ Suspicious finance/social media messages
Social Engineering β†’ Manipulative messages exploiting trust

πŸ”œ Next Steps
- Expanding datasets with more phishing & malware examples
- Refining schema & annotation quality
- Open to feedback, contributions, and suggestions

If this is something you might find useful, bookmark/follow/like the dataset repo <3

πŸ’¬ Thoughts, feedback, and ideas are always welcome! Drop a comment or DMs are open πŸ€™
reacted to their post with πŸ‘€ 24 days ago
view post
Post
1398
So, what is #MechanisticInterpretability πŸ€”

Mechanistic Interpretability (MI) is the discipline of opening the black box of large language models (and other neural networks) to understand the underlying circuits, features and/or mechanisms that give rise to specific behaviours

Instead of treating a model as a monolithic function, we can:

1. Trace how input tokens propagate through attention heads & MLP layers
2. Identify localized β€œcircuit motifs”
3. Develop methods to systematically break down or β€œedit” these circuits to confirm we understand the causal structure.

Mechanistic Interpretability aims to yield human-understandable explanations of how advanced models represent and manipulate concepts which hopefully leads to

1. Trust & Reliability
2. Safety & Alignment
3. Better Debugging / Development Insights

https://bsky.app/profile/mechanistics.bsky.social/post/3lgvvv72uls2x
  • 1 reply
Β·
posted an update 24 days ago
view post
Post
1398
So, what is #MechanisticInterpretability πŸ€”

Mechanistic Interpretability (MI) is the discipline of opening the black box of large language models (and other neural networks) to understand the underlying circuits, features and/or mechanisms that give rise to specific behaviours

Instead of treating a model as a monolithic function, we can:

1. Trace how input tokens propagate through attention heads & MLP layers
2. Identify localized β€œcircuit motifs”
3. Develop methods to systematically break down or β€œedit” these circuits to confirm we understand the causal structure.

Mechanistic Interpretability aims to yield human-understandable explanations of how advanced models represent and manipulate concepts which hopefully leads to

1. Trust & Reliability
2. Safety & Alignment
3. Better Debugging / Development Insights

https://bsky.app/profile/mechanistics.bsky.social/post/3lgvvv72uls2x
  • 1 reply
Β·