Philipp Schmid 3/23/2023

Efficient Large Language Model training with LoRA and Hugging Face

Read Original

This tutorial demonstrates how to apply Parameter Efficient Fine-Tuning (PEFT), specifically the LoRA technique, to fine-tune the 11-billion parameter FLAN-T5 XXL model for a summarization task using the samsum dataset. It provides a step-by-step guide covering environment setup with Hugging Face Transformers, Accelerate, and PEFT, dataset preparation, model training with LoRA and bitsandbytes int-8 quantization, and subsequent evaluation and inference.

Efficient Large Language Model training with LoRA and Hugging Face

Comments

No comments yet

Be the first to share your thoughts!

Browser Extension

Get instant access to AllDevBlogs from your browser