Ahead of AI #12: LLM Businesses and Busyness
Darren LI and added
sari and added
- The best LLMs are going to be much better in all the different dimensions you probably care about (including e.g. less hallucinations)
- The best open-sourced LLMs are going to be better than the best non-open source LLMs nowadays. As an example, Facebook AI made a big announcement for their LLaMA open source model a couple of weeks back . A few days
Towards Reinforcement Learning with AI Feedback (RLAIF). What open-sourced foundation models, instruction tuning, and other recent events mean for the future of AI
Nicolay Gerold added
As the LLM market structure stabilizes, the next frontier is now emerging. The focus is shifting to the development and scaling of the reasoning layer, where “System 2” thinking takes precedence. Inspired by models like AlphaGo, this layer aims to endow AI systems with deliberate reasoning, problem-solving and cognitive operations at inference time... See more
Sonya Huang • Generative AI’s Act O1
In 2023, AI must absorb an influx of tourists – thin GPT wrappers, MBA tweet threads about AI trends, LinkedIn bios changing from #crypto to #AI. Genuine technologists will join the development effort, but filtering out negative human capital is challenging.
Undoubtedly, more builders are needed to apply and productionize the latest technologies,
... See moreJohn Luttig • Is AI the new crypto?
sari and added
Amplify Partners was running a survey among 800+ AI engineers to bring transparency to the AI Engineering space. The report is concise, yet it provides a wealth of insights into the technologies and methods employed by companies for the implementation of AI products.
Highlights
👉 Top AI use cases are code intelligence, data extraction and workflow a... See more
Highlights
👉 Top AI use cases are code intelligence, data extraction and workflow a... See more
Feed | LinkedIn
pair-preference-model-LLaMA3-8B by RLHFlow: Really strong reward model, trained to take in two inputs at once, which is the top open reward model on RewardBench (beating one of Cohere’s).
DeepSeek-V2 by deepseek-ai (21B active, 236B total param.): Another strong MoE base model from the DeepSeek team. Some people are questioning the very high MMLU sc... See more
DeepSeek-V2 by deepseek-ai (21B active, 236B total param.): Another strong MoE base model from the DeepSeek team. Some people are questioning the very high MMLU sc... See more
Shortwave — rajhesh.panchanadhan@gmail.com [Gmail alternative]
Nicolay Gerold added