Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

If you want to cut costs, I would recommend you try experimenting with Meta’s new Llama 3 model


I actually played with groq / llama3 earlier this week! It is super fast. I got into the situation where the 70b model is better than haiku, but more expensive and the 8B model was significantly worse. I still think haiku is great bang / buck ratio atm. I want to try tuning the prompts more for llama3 though, I'm sure I can get comparable performance with enough effort.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: