JP van Oosten

Training a language model on a single GPU in one day

Feb 2, 2023

The other day, I mentioned nanoGPT in one of my posts, an implementation of GPT that smaller organisations can use for generating text. This type of research is cool because it creates more diversity of use of these types of models. I’ll leave my thoughts about going all in on the "deep-learning bet" for another time 😉

In December, two researchers from the University of Maryland published a pre-print on “Cramming”, a challenge they made to see how much of a BERT-style language model can be trained in one day, on one GPU. They use a number of tricks to make the model smaller and more performant for a specific task.

The paper shows that they got pretty far in terms of performance. Even given the limitations of this smaller model, this type of research is inspiring, because it will allow individuals and small organisations to also use language modelling successfully.

Of course, most of the time, you don’t need to train your own BERT-like models, and can just use pre-trained models. Or, you might not even need fancy deep learning models for solving your business problems. Are you wondering what kind of AI technology is best for your business? Feel free to drop me a line and we’ll have a chat!

(Also posted on my LinkedIn feed)