Русские видео

Сейчас в тренде

Иностранные видео


Скачать с ютуб What is Low-Rank Adaptation (LoRA) | explained by the inventor в хорошем качестве

What is Low-Rank Adaptation (LoRA) | explained by the inventor 9 месяцев назад


Если кнопки скачивания не загрузились НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если возникают проблемы со скачиванием, пожалуйста напишите в поддержку по адресу внизу страницы.
Спасибо за использование сервиса savevideohd.ru



What is Low-Rank Adaptation (LoRA) | explained by the inventor

Low-rank adaptation, or LoRA, is one of the most popular methods for customizing large AI models. Why was it invented? What is it? Why should I consider using it? Find out the answers in this video from the inventor of LoRA. *Like, subscribe, and share if you find this video valuable!* Paper: https://arxiv.org/abs/2106.09685 Repo: https://github.com/microsoft/lora 0:00 - Intro 0:34 - How we came up with LoRA 1:33 - What is LoRA? 3:14 - How to choose the rank r? 3:57 - Does LoRA work for my model architecture? 4:48 - Benefits of using LoRA 6:03 - Engineering ideas enabled by LoRA Training or serving multiple LoRA modules in a single batch with the following community implementations: https://github.com/S-LoRA/S-LoRA https://github.com/sabetAI/BLoRA https://github.com/TUDB-Labs/multi-lo... Follow me on Twitter:   / edwardjhu   🙏Gratitude: Super grateful to my coauthors on the LoRA paper: Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen.

Comments