Today I'm sharing this 🔥 video from our recent Engineering onsite at Demandbase HQ in SF! https://lnkd.in/eFAFR6QG We were so fortunate to host Zhengxuan Wu and Aryaman Arora, two upcoming AI researchers from Stanford University Department of Computer Science. Their approach to fine tuning, called ReFT, targets the representations emitted by neural net layers rather than the weights of the layers themselves. This makes it very flexible when using the same base model for swapping multiple fine tunes. The memory overhead is also very low for their intervention weights. Another highlight is how powerful it can be. One POC they showed forced the LLM to speak in only emojis exhibiting ReFT's power to control output. Control is crucial when, for example, you want to prevent your bot from offering customers unauthorized refunds!! If you're already familiar with LoRA, it should be pretty simple to use with only a few extra hyperparameters -- details in the video. Here's a link to the repo. Give them a star! https://lnkd.in/eJScNbzK Thanks again for letting us host your first external talk Aryaman and Zen. It's very exciting tech!!
ReFT: Representation Finetuning for Language Models -- Aryaman Arora & Zhengxuan (Zen) Wu
https://www.youtube.com/
This is very cool, Josh. Thanks for sharing!
This is awesome!