[TOC]
Ref. from Yann Leccun video
Figure by UbiquitousLearning/Efficient LLM and Foundation Models
Picture from: Training Large Foundation Models Using SageMaker HyperPod by Ian Gibbs - Senior PMT-ES in AI/ML - Gen AI Enablement Weekly Series
By IEEE Spectrum https://spectrum.ieee.org/ai-index-2024
Guide to when to prompt versus fine tuning considering different organizations?
PEFT, or Parameter Efficient Fine-tuning, is a Hugging Face open-source library to enable efficient adaptation of pre-trained language models (PLMs) to various downstream applications without fine-tuning all the model’s parameters. PEFT currently includes techniques for:
Prefix Tuning: P-Tuning v2: Prompt Tuning Can Be Comparable to Fine-tuning Universally Across Scales and Tasks
P-Tuning: GPT Understands, Too
Prompt Tuning: The Power of Scale for Parameter-Efficient Prompt Tuning
IA3: Infused Adapter by Inhibiting and Amplifying Inner Activations
Reference:
[1] LoRA Serving on Amazon SageMaker — Serve 100’s of Fine-Tuned LLMs For the Price of 1
[3] Github example: Fine-tune LLaMA 2 on Amazon SageMaker
[4] GitHub example: Fine-tune LLaMA 2 models on SageMaker JumpStart
[6] Ref.scaling down to scale up a guide to parameter-efficient fine-tuning
[7] LoRA Paper
[8] LoRA Land: Fine-Tuned Open-Source LLMs that Outperform GPT-4
Ref. The Ultimate Battle of Language Models: Lit-LLaMA vs GPT3.5 vs Bloom vs …
Proximal Policy Optimization (PPO): reinforcement learing algorithm
Batch size:
number of experiences sampled at random from an experience buffer and used to update the neural network weights.
The batch is a subset of an experience buffer that is composed of images captured by the camera mounted on the AWS DeepRacer vehicle and actions taken by the vehicle.
Epochs:
Learning rate:
Entropy:
Discount Factor:
Loss type:
Responsible Generative AI: A Code of Ethics for the Future