Sublime
An inspiration engine for ideas

While Apple has been positioning M4 chips for local AI inference with their unified memory architecture, NVIDIA just undercut them massively.
Stacking Project Digits personal computers is now the most affordable way to run frontier LLMs locally.
The 1 petaflop headline feels like marketing h... See more




People are just starting to realize the power of Local LLMs
Especially with new Apple chips. It's a game-changer
Let me show you:
Falcon is 180B LLM. It is the size of GPT3.
How fast would it run on your Mac?
Apple M3 Max: 3.5 tokens/sec___LINEBREAK... See more

$QUIK
Quick and easy to understand what the latest developments in $INTC mean for $QUIK
The oppurtunity here is being slept on imo
With the aid of Perplexity
QuickLogic: The Sole eFPGA Powerhouse on Intel 18A Poised for Explosive Grow... See more
TheUndefinedMysticx.com

Chat, @deepseek_ai just dropped DeepGEMM š„
> a lightweight CUDA-based library for efficient FP8 GEMMs on NVIDIA Hopper tensor cores
> featuring JIT compilation, CUDA-core two-level accumulation
> ~300-line core kernel
> outperforming expert-tuned libraries like CUTLASS 3.6 in D... See more