Generative AI with Large Language Models

Coursera - week2 Fine-tuning LLMs

Posted by Xuan on July 10, 2023

Week 1 - Fine-tuning LLMs with instruction

In contrast to pre-training, where you train the LLM via self-supervised learning, Fine-tuning is a supervised learning process where you used a data set of labeled examples to update the weights of the LLMs

  • Full fine-tuning (updates all parameters)
  • Instruction fine-tuning

parameter efficient fine-tuning or PEFT

PEFT is a set of techniques that preserves the weights of the original LLM and trains only a small number of task-specific adapter layers and parameters. PEFT shows greater robustness to catastrophic forgetting since most of the pre-trained weights are left unchanged. PEFT is an exciting and active area of research that we will cover later this week. In the meantime, let’s move on to the next video and take a closer look at multitask fine-tuning.

Fine-tuning for multitasks

FLAN family (Fine-tuned LAnguage Net)

FLAN, which stands for fine-tuned language net, is a specific set of instructions used to fine-tune different models. FLAN

LoRA

because of the performance results of using those low rank matrices as opposed to full fine-tuning