quickly train llama-7b on some data

#large_language_models #notes #author_luna

i was able to use https://github.com/Lightning-AI/lit-llama to finetune llama-7b on a dual-3090 setup from a colleague of mine

however, the project does not support unstructured datasets, so i hacked together an implementation of such and posted it out here: https://github.com/Lightning-AI/lit-llama/pull/278 so you can use it directly (by appending .patch to the url, and git applying the result of that)