Check out the article & Follow, bookmark, save the tab as I will be updating it <3 (using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
Check out the article & Follow, bookmark, save the tab as I will be updating it <3 (using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
Check out the article & Follow, bookmark, save the tab as I will be updating it <3 (using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
Fine-tuning your LLM is like min-maxing your ARPG hero so you can push high-level dungeons and get the most out of your build/gear... Makes sense, right? π
Here's a cheat sheet for devs (but open to anyone!)
---
TL;DR
- Full Fine-Tuning: Max performance, high resource needs, best reliability. - PEFT: Efficient, cost-effective, mainstream, enhanced by AutoML. - Instruction Fine-Tuning: Ideal for command-following AI, often combined with RLHF and CoT. - RAFT: Best for fact-grounded models with dynamic retrieval. - RLHF: Produces ethical, high-quality conversational AI, but expensive.
Choose wisely and match your approach to your task, budget, and deployment constraints.
I just posted the full extended article here if you want to continue reading >>>
Fine-tuning your LLM is like min-maxing your ARPG hero so you can push high-level dungeons and get the most out of your build/gear... Makes sense, right? π
Here's a cheat sheet for devs (but open to anyone!)
---
TL;DR
- Full Fine-Tuning: Max performance, high resource needs, best reliability. - PEFT: Efficient, cost-effective, mainstream, enhanced by AutoML. - Instruction Fine-Tuning: Ideal for command-following AI, often combined with RLHF and CoT. - RAFT: Best for fact-grounded models with dynamic retrieval. - RLHF: Produces ethical, high-quality conversational AI, but expensive.
Choose wisely and match your approach to your task, budget, and deployment constraints.
I just posted the full extended article here if you want to continue reading >>>
π Whatβs in v0.1? A few structured scam examples (text-based) Covers DeFi, crypto, phishing, and social engineering Initial labelling format for scam classification
β οΈ This is not a full dataset yet (samples are currently available). Just establishing the structure + getting feedback.
π Current Schema & Labelling Approach "instruction" β Task prompt (e.g., "Evaluate this message for scams") "input" β Source & message details (e.g., Telegram post, Tweet) "output" β Scam classification & risk indicators
ποΈ Current v0.1 Sample Categories Crypto Scams β Meme token pump & dumps, fake DeFi projects Phishing β Suspicious finance/social media messages Social Engineering β Manipulative messages exploiting trust
π Next Steps - Expanding datasets with more phishing & malware examples - Refining schema & annotation quality - Open to feedback, contributions, and suggestions
If this is something you might find useful, bookmark/follow/like the dataset repo <3
π¬ Thoughts, feedback, and ideas are always welcome! Drop a comment or DMs are open π€
π Whatβs in v0.1? A few structured scam examples (text-based) Covers DeFi, crypto, phishing, and social engineering Initial labelling format for scam classification
β οΈ This is not a full dataset yet (samples are currently available). Just establishing the structure + getting feedback.
π Current Schema & Labelling Approach "instruction" β Task prompt (e.g., "Evaluate this message for scams") "input" β Source & message details (e.g., Telegram post, Tweet) "output" β Scam classification & risk indicators
ποΈ Current v0.1 Sample Categories Crypto Scams β Meme token pump & dumps, fake DeFi projects Phishing β Suspicious finance/social media messages Social Engineering β Manipulative messages exploiting trust
π Next Steps - Expanding datasets with more phishing & malware examples - Refining schema & annotation quality - Open to feedback, contributions, and suggestions
If this is something you might find useful, bookmark/follow/like the dataset repo <3
π¬ Thoughts, feedback, and ideas are always welcome! Drop a comment or DMs are open π€
Mechanistic Interpretability (MI) is the discipline of opening the black box of large language models (and other neural networks) to understand the underlying circuits, features and/or mechanisms that give rise to specific behaviours
Instead of treating a model as a monolithic function, we can:
1. Trace how input tokens propagate through attention heads & MLP layers 2. Identify localized βcircuit motifsβ 3. Develop methods to systematically break down or βeditβ these circuits to confirm we understand the causal structure.
Mechanistic Interpretability aims to yield human-understandable explanations of how advanced models represent and manipulate concepts which hopefully leads to