ttkciar@alien.topB to LocalLLaMAEnglish · 2 years agoPractical Tips for Finetuning LLMs Using LoRA (Low-Rank Adaptation)magazine.sebastianraschka.comexternal-linkmessage-square1linkfedilinkarrow-up11arrow-down10
arrow-up11arrow-down1external-linkPractical Tips for Finetuning LLMs Using LoRA (Low-Rank Adaptation)magazine.sebastianraschka.comttkciar@alien.topB to LocalLLaMAEnglish · 2 years agomessage-square1linkfedilink
minus-squareRelevant_Outcome_726@alien.topBlinkfedilinkEnglisharrow-up1·2 years agoFrom my experience, Here are some other things related to Lora: + FSDP doesn’t work for Lora because FSDP requires all parameters to be trainable or frozen. + For Qlora, currently we can only use deepspeed zero2 (deepspeed zero 3 is not supported)
From my experience, Here are some other things related to Lora:
+ FSDP doesn’t work for Lora because FSDP requires all parameters to be trainable or frozen.
+ For Qlora, currently we can only use deepspeed zero2 (deepspeed zero 3 is not supported)