Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Someone brought up an interesting point: to get the latest data (news, scientific breakthroughs...) into the model, you need to constantly retrain it.


The incremental compute costs will scale with the incremental data added, therefore training costs will grow at a much slower rate compared to when training was GPU limited.


Or, you know, use rag. Which is far better and more accurate than regurgitating compressed training knowledge.


Oh please




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: