During the process of writing AI Engineering, I went through so many papers, case studies, blog posts, repos, tools, etc. This repo contains ~100 resources that really helped me understand various aspects of building with foundation models.
https://t.co/cmfZw5LgiX
Here are the highlights:
1. Anthropic’s Prompt Engineering Interactive Tutorial
The Google Sheets-based interactive exercises make it easy to experiment with different prompts and see immediately what works and what doesn’t. I’m surprised other model providers don’t have similar interactive guides: https://t.co/D0IP1eCMOI
2. OpenAI’s best practices for finetuning
While this guide focuses on GPT-3, many techniques are applicable to full finetuning in general. It explains how finetuning works, how to prepare training data, how to pick training hyperparameters, and common finetuning mistakes: https://t.co/JENqR2V17I
3. Llama 3 paper
The section on post-training data is a gold mine as it details different techniques they used to generate 2.7 million examples for supervised finetuning. It also covers a crucial but less talked about topic: data verification, how to evaluate the quality of synthetic data: https://t.co/n0ysigxy9q
4. Efficiently Scaling Transformer Inference (Pope et al., 2022)
An amazing paper co-authored by Jeff Dean about inference optimization for transformers models. It covers not only different optimization techniques and their tradeoffs, but also provides a guideline for what to do if you want to optimize for different aspects, e.g. lowest possible latency, highest possible throughput, or longest context length: https://t.co/hyKsp1QY2A
5. Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models (Lu et al., 2023)
My favorite study on LLM planners, how they use tools, and their failure modes. An interesting finding is that different LLMs have different tool preferences: https://t.co/cqRAlu9yOr
6.... See more
Chip Huyenx.com