How to Add Special Tokens to LLMs Safely
Learn how to add special tokens to LLMs during fine-tuning without causing catastrophic forgetting. Our guide covers smart initialization and PEFT/LoRA.
Bao Bao Suan Fa Bi Ji
Dive deep into the world of Artificial Intelligence with our curated collection of articles, covering the latest breakthroughs and insights from leading researchers and engineers.
Learn how to add special tokens to LLMs during fine-tuning without causing catastrophic forgetting. Our guide covers smart initialization and PEFT/LoRA.
Bao Bao Suan Fa Bi Ji
Fine-tune LLaMA 3 with zero coding in 3 steps using LLaMA Factory WebUI. Save 80% GPU memory with QLoRA on RTX 3090/4090. Beginner-friendly tutorial with CUDA setup. Supports 100+ models.
Number in the Moutain
Discover a critical flaw in Supervised Fine-Tuning (SFT) that limits LLM performance. Learn how a simple learning rate tweak unifies SFT and DPO for a 25% gain.
Alex
Master Supervised Fine-Tuning (SFT) transforming base models to chat assistants. Complete 3-stage pipeline: base → instruct → chat model. LoRA reduces cost 70%, 7B model SFT in 2-4 hours on A100 ($10-20). Alpaca vs Dolly vs Open-Orca datasets compared.
Alex