Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

It does start getting worse at some point right?


Cross-entropy loss can start getting worse due to the model becoming less calibrated, even as the classification accuracy continues to improve. I first heard that here: https://arxiv.org/abs/1706.04599

Is this 'overconfidence' the leading explanation as to why LLMs continue to show qualitative improvement even after their test loss levels off?


Is it possible to somehow modify the sampling from the model to account for that?


I'm sure eventually it would, but we haven't gotten to that point yet in our training.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: