Check out the article & Follow, bookmark, save the tab as I will be updating it <3 (using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
Check out the article & Follow, bookmark, save the tab as I will be updating it <3 (using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
Check out the article & Follow, bookmark, save the tab as I will be updating it <3 (using it as my own notepad & decided i might keep it up to date if i post it here, instead of making the 15th_version of it and not saving it with a name i can remember on my desktop lol)
Fine-tuning your LLM is like min-maxing your ARPG hero so you can push high-level dungeons and get the most out of your build/gear... Makes sense, right? π
Here's a cheat sheet for devs (but open to anyone!)
---
TL;DR
- Full Fine-Tuning: Max performance, high resource needs, best reliability. - PEFT: Efficient, cost-effective, mainstream, enhanced by AutoML. - Instruction Fine-Tuning: Ideal for command-following AI, often combined with RLHF and CoT. - RAFT: Best for fact-grounded models with dynamic retrieval. - RLHF: Produces ethical, high-quality conversational AI, but expensive.
Choose wisely and match your approach to your task, budget, and deployment constraints.
I just posted the full extended article here if you want to continue reading >>>
Fine-tuning your LLM is like min-maxing your ARPG hero so you can push high-level dungeons and get the most out of your build/gear... Makes sense, right? π
Here's a cheat sheet for devs (but open to anyone!)
---
TL;DR
- Full Fine-Tuning: Max performance, high resource needs, best reliability. - PEFT: Efficient, cost-effective, mainstream, enhanced by AutoML. - Instruction Fine-Tuning: Ideal for command-following AI, often combined with RLHF and CoT. - RAFT: Best for fact-grounded models with dynamic retrieval. - RLHF: Produces ethical, high-quality conversational AI, but expensive.
Choose wisely and match your approach to your task, budget, and deployment constraints.
I just posted the full extended article here if you want to continue reading >>>
π Whatβs in v0.1? A few structured scam examples (text-based) Covers DeFi, crypto, phishing, and social engineering Initial labelling format for scam classification
β οΈ This is not a full dataset yet (samples are currently available). Just establishing the structure + getting feedback.
π Current Schema & Labelling Approach "instruction" β Task prompt (e.g., "Evaluate this message for scams") "input" β Source & message details (e.g., Telegram post, Tweet) "output" β Scam classification & risk indicators
ποΈ Current v0.1 Sample Categories Crypto Scams β Meme token pump & dumps, fake DeFi projects Phishing β Suspicious finance/social media messages Social Engineering β Manipulative messages exploiting trust
π Next Steps - Expanding datasets with more phishing & malware examples - Refining schema & annotation quality - Open to feedback, contributions, and suggestions
If this is something you might find useful, bookmark/follow/like the dataset repo <3
π¬ Thoughts, feedback, and ideas are always welcome! Drop a comment or DMs are open π€