GitHub - okuvshynov/slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization

GitHub - okuvshynov/slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization

okuvshynovgithub.com
Thumbnail of GitHub - okuvshynov/slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization

Lightning-AI GitHub - Lightning-AI/litgpt: Pretrain, finetune, deploy 20+ LLMs on your own data. Uses state-of-the-art techniques: flash attention, FSDP, 4-bit, LoRA, and more.

mit-han-lab GitHub - mit-han-lab/streaming-llm: Efficient Streaming Language Models with Attention Sinks

Thumbnail of www-x-com-karpathy-status-1811467135279104217
Thumbnail of www-twitter-com-akhaliq-status-1737300118070534468

GitHub - mit-han-lab/streaming-llm: Efficient Streaming Language Models with Attention Sinks

mit-han-labgithub.com
Thumbnail of GitHub - mit-han-lab/streaming-llm: Efficient Streaming Language Models with Attention Sinks

Llama 2 - Resource Overview - Meta AI

ai.meta.com
Thumbnail of Llama 2 - Resource Overview - Meta AI

How I run LLMs locally

abishekmuthian.com
Cover of How I run LLMs locally