So my LLM LLama2 training did work

 So it took about 4 hours on my mac without GPU improvements but this command ran for about 4 hours.

python3 train.py --compile=False --vocab_source=custom --vocab_size=4096


And then I could run against the llama2.c code.

 ./run ./out/model.bin

I got this gibberish:

"pres es eluted эargared copy É Int beforepperunch          KarologfromwayClassistoryork ochidentAr}^ Allet Com easgoogleiden targetegaoman essпиgesscript non behS commandasesasesба amb before pervecnode agcolorkoeln conf Ma Setrat
Textema governiowhere ##скогоchange.) respectankön knowPar namesiones неander für enrid muool medcia depBalityви rangehelова () del options ### voando](arget Thereise und descri L`,incless++readble oldredULLockabelutesphaires says буClientIC});viroo test only ser"

Comments

Popular posts from this blog

JVM Notebook: Basic Clojure, Java and JVM Language performance

On Unit Testing, Java TDD for developers to write

Is Java the new COBOL? Yes. What does that mean, exactly? (Part 1)