Why is Discord such a good GTM for AI applications?
Text interface. Most users are just generating images, videos, and audio in these Discord servers. Prompts are easily expressible in simple text commands. It’s why we’ve seen image generation strategies like Midjourney (all-in-one) flourish in Discord while more raw diffusion models haven’t grown... See more
The main thing Data Science forgot is that understanding the data and the goals for the data are essential. In doing so, they fundamentally forgot their roots when it comes to Operations Research and Statistics which have a fundamental rule each:
Operations Research - Answering the why before the what or how.
Statistics - Form a null hypothesis and... See more
It’s underdigitized. According to the McKinsey Industry Digitalization Index, only the agriculture sector is less digitized than construction. The typical IT spend for construction companies is 1-2% of the revenue, compared with the 3-5% average across industries. Moreover, there are many barriers to digital technology adoption including skill... See more
Alex Albert shared 5 architectures for using them in an agentic context:
Delegation : Use cheaper, faster models for cost and speed gains.
For example, Opus can delegate to Haiku to read a book and return relevant passages. This works well if the task description & result are more compact than the full context.
What’s the best way for an end user to organize and explore millions of latent space features?
I’ve found tens of thousands of interpretable features in my experiments, and frontier labs have demonstrated results with a thousand times more features in production-scale models. No doubt, as interpretability techniques advance, we’ll see feature maps... See more
The LLM doesn’t call the tool directly (yet), but it does pass back to the application what functions should be called — and with which parameters. And, now, OpenAI lets multiple function calls be “invoked” at once.
But, this idea is not just about GPT. The open source world is moving towards this model as well.
Advances in data processing techniques. You can increase context length in two ways. First, you can train the model with longer context lengths. That’s difficult because it’s much more computationally expensive, and it’s hard to find datasets with long context lengths (most documents in CommonCrawl have fewer than 2,000 tokens).