Entry 08 / 09
Mar 2024
Llama2 Fine-tuning
End-to-end pipeline for fine-tuning Llama2 large language models using LoRA / PEFT adapters. Covers data preparation, supervised fine-tuning, and evaluation - built as a reproducible notebook for experimenting with domain adaptation of open-weight LLMs.
PythonLlama2LoRAPEFTJupyter
A learning project to internalize the LoRA pipeline end-to-end: data prep, supervised fine-tuning, adapter merge, and a small eval harness. Scope was deliberately narrow so I could touch every piece. The base model is dated now, but the muscle memory transferred - when I reach for PEFT in newer work, it's because of what this project taught me about adapters and merge mechanics.