Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I'm sorry, but why would you run a model like ALBERT on a CPU in the first place?

It's pretty common to run inference on CPUs. There are lots of operational and cost reasons why this makes sense in at least some cases.



I would even say: in every case, except where a GPU or TPU is necessary to achieve a certain speed. Unless there are very specific reasons for it, GPU/TPU is just unnecessarily cost inefficient.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: