PEFT-FLANT5
2022In this notebook, we will fine-tune an existing LLM from Hugging Face for enhanced dialogue summarization. we will use the FLAN-T5 model, which provides a high quality instruction tuned model and can summarize text out of the box. To improve the inferences, we will explore a full fine-tuning approach and evaluate the results with ROUGE metrics. Then we will perform Parameter Efficient Fine-Tuning (PEFT), evaluate the resulting model and see that the benefits of PEFT outweigh the slightly-lower performance metrics.
- Notebook
- Notebook
- Platform
- Google Colab
- Stack
- PyTorch, Transformers, PEFT