Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
How "Hardwired" AI Will Destroy Nvidia's Empire and Change the World (medium.com/mokrasar)
12 points by amelius 3 hours ago | hide | past | favorite | 8 comments
 help



The foundation models themselves will be cheap to deploy, but we’ll still need general purpose inferencing hardware to work along side them, converting latent intermediate layers to useful, application-specific concerns. This may level off the demand for “gpu/tpu” hardware, though, by letting the biggest and most expensive layers move to silicon.

This is still far away from being viable for actually useful models, like bigger MoE ones with much larger context windows. I mean, the technology is very promising just like Cerebras, but we need to see whether they are able to keep up this with the evolution of the models to come in the next few years. Extremely interesting nevertheless.

Is this a paid ad placement? I'm seeing a load of breathless "commentary" on Taalas and next to no serious discussion about whether their approach is even remotely scalable. A one-off tech demo using a comparatively ancient open source model is hardly going to be giving Jensen Huang sleepless nights.

I always thought once we have the models figured out, getting the meat of it into an FPGA was probably the logical next step. They seemed to have skipped that and are directly writing the program as ASIC (ROM). Pretty wild.

It's crazy. In a few years we will be able to buy Qwen on a chip, doing 10K tokens per second.

Yeah, well might just come on your new laptop

Or your phone.

Give me a 120B dense model on one of these and yeah my API use will probably drop.



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: