So my LLM LLama2 training did work
So it took about 4 hours on my mac without GPU improvements but this command ran for about 4 hours.
python3 train.py --compile=False --vocab_source=custom --vocab_size=4096
And then I could run against the llama2.c code.
./run ./out/model.bin
I got this gibberish:
"pres es eluted эargared copy É Int beforepperunch KarologfromwayClassistoryork ochidentAr}^ Allet Com easgoogleiden targetegaoman essпиgesscript non behS commandasesasesба amb before pervecnode agcolorkoeln conf Ma Setrat
Textema governiowhere ##скогоchange.) respectankön knowPar namesiones неander für enrid muool medcia depBalityви rangehelова () del options ### voando](arget Thereise und descri L`,incless++readble oldredULLockabelutesphaires says буClientIC});viroo test only ser"
Comments