quickly train llama-7b on some data
#large_language_models #notes #author_luna
i was able to use https://github.com/Lightning-AI/lit-llama to finetune llama-7b on a dual-3090 setup from a colleague of mine
however, the project does not support unstructured datasets, so i hacked together an implementation of such and posted it out here: https://github.com/Lightning-AI/lit-llama/pull/278 so you can use it directly (by appending .patch
to the url, and git apply
ing the result of that)
all of that was merged