
Training A Not-So-Large Language Model for $50
September 08, 2023 | 3 min readWhat I cannot create, I do not understand. Let's train your own LLM!
Under the sea, in the hippocampus's garden...
What I cannot create, I do not understand. Let's train your own LLM!
A quick guide for RLHF using trlX, OPT-1.5B, and LoRA.
Transformer has undergone various application studies, model enhancements, etc. This post aims to provide an overview of these studies.
This post explains how MobileBERT succeeded in reducing both model size and inference time and introduce its implementation in TensorFlow.js that works on web browsers.