Sublime
An inspiration engine for ideas
on-demand H100 for $0.99/hr, 4090 for $0.20/hr at Hyperbolic
likely the cheapest GPUs around
tell me what you're building, and I'll spot you free credits for an 8xH100 node for at least a few hours to start. https://t.co/SIwpzrZ0gL
Yuchen Jinx.com

I hate to acknowledge this, but Gemini 1.5 Flash is better than llama-3-70b on long context tasks. It's way faster than my locally hosted 70b model (on 4*A6000) and hallucinates less. The free of charge plan is good enough for me to do prompt engineering for prototyping https://t.co/pJJMvScbhj
Diffusion language models are SO FAST!!
A new startup, Inception Labs, has released Mercury Coder, "the first commercial-scale diffusion large language model"
It's 5-10x faster than current gen LLMs, providing high-quality responses at low costs.
And you can try... See more
Tanishq Abraham is at ICMLx.comBREAKING: DeepSeek released R1.
And it's already better than ChatGPT o1.
1. It's open-source.
2. API is 96.4% cheaper than chatgpt.
3. I run my tests below. A quick thread: https://t.co/U3nj4HZYUu
Ruben Hassidx.com
Andrej Karpathy released nanochat, ~8K lines of minimal code that do pretrain + midtrain + SFT + RL + inference + ChatGPT-like webUI.
It trains a 560M LLM in ~4 hrs on 8×H100.
I trained and hosted it on Hyperbolic GPUs ($48). First prompt reminded me how funny tiny LLMs... See more

happy saturday
for the price of a cup of coffee, you can rent a 4090 for the day and do some random experiments on small models https://t.co/5yGBU7J6re
introducing rabbitOS intern
put the new intern to work at https://t.co/R3sOtVWoJ5 today (for free) https://t.co/jCfAMcIJA0
rabbit inc.x.com