darkz baron

darkzbaron

AI & ML interests

nlp and cv and ai

Recent Activity

liked a model about 17 hours ago
deepseek-ai/DeepSeek-R1-Zero
liked a Space about 18 hours ago
yeliudev/VideoMind-2B
View all activity

Organizations

None yet

darkzbaron's activity

reacted to odellus's post with ๐Ÿ‘ 3 days ago
view post
Post
1881
Super grateful to @marriola for the release of the block diffusion code and model. I'm generating text with diffusion locally! Couldn't be more pleased.
  • 2 replies
ยท
reacted to onekq's post with ๐Ÿš€ 6 days ago
view post
Post
2548
We desperately need GPU for model inference. CPU can't replace GPU.

I will start with the basics. GPU is designed to serve predictable workloads with many parallel units (pixels, tensors, tokens). So a GPU allocates as much transistor budget as possible to build thousands of compute units (Cuda cores in NVidia or execution units in Apple Silicon), each capable of running a thread.

But CPU is designed to handle all kinds of workloads. CPU cores are much larger (hence a lot fewer) with branch prediction and other complex things. In addition, more and more transistors are allocated to build larger cache (~50% now) to house the unpredictable, devouring the compute budget.

Generalists can't beat specialists.
ยท