Running 10 EmbeddingGemma Tuning Lab ๐ป 10 Fine-tune EmbeddingGemma to understand your personal taste
view post Post 1615 Are you familiar with reverse residual connections or looping in language models?Excited to share my Looped-GPT blog post and codebase ๐https://github.com/sanyalsunny111/Looped-GPTTL;DR: looping during pre-training improves generalization.Plot shows GPT2 LMs pre-trained with 15.73B OWT tokensP.S. This is my first post here โ I have ~4 followers and zero expectations for reach ๐ See translation 3 replies ยท ๐ง 6 6 ๐ 3 3 + Reply