GitHub - okuvshynov/slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization

GitHub - okuvshynov/slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization

okuvshynovgithub.com
Thumbnail of GitHub - okuvshynov/slowllama: Finetune llama2-70b and codellama on MacBook Air without quantization

Llama 2 - Resource Overview - Meta AI

ai.meta.com
Thumbnail of Llama 2 - Resource Overview - Meta AI

Discover, Download, and Run Local LLMs

lmstudio.ailmstudio.ai
Thumbnail of Discover, Download, and Run Local LLMs

turboderp GitHub - turboderp/exllamav2: A fast inference library for running LLMs locally on modern consumer-class GPUs

GitHub - mit-han-lab/streaming-llm: Efficient Streaming Language Models with Attention Sinks

mit-han-labgithub.com
Thumbnail of GitHub - mit-han-lab/streaming-llm: Efficient Streaming Language Models with Attention Sinks

jmorganca GitHub - jmorganca/ollama: Get up and running with Llama 2 and other large language models locally

Mozilla-Ocho GitHub - Mozilla-Ocho/llamafile: Distribute and run LLMs with a single file.

unslothai GitHub - unslothai/unsloth: 5X faster 50% less memory LLM finetuning