GitHub - okuvshynov/slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization

We made a Guide to teach you how to Fine-tune LLMs correctly!
Learn about:
• Choosing the right parameters & training method
• RL, GRPO, DPO & CPT
• Data prep, Overfitting & Evaluation
• Training with Unsloth & deploy on vLLM, Ollama, Open WebUI___L... See more
GitHub - mit-han-lab/streaming-llm: Efficient Streaming Language Models with Attention Sinks
mit-han-labgithub.comOops haven't tweeted too much recently; I'm mostly watching with interest the open source LLM ecosystem experiencing early signs of a cambrian explosion. Roughly speaking the story as of now:
1. Pretraining LLM base models remains very expensive. Think: supercomputer + months.
2. But finetuning LLMs is tur... See more
Andrej Karpathyx.com