GitHub - young-geng/EasyLM: Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.
Tinker is cool.
If you're a researcher/developer, tinker dramatically simplifies LLM post-training. You retain 90% of algorithmic creative control (usually related to data, loss function, the algorithm) while tinker handles the hard parts that you usually want to touch much less often (infra, forward/backward of the LLM... See more
Andrej Karpathyx.comOverview
MaxText is a high performance , highly scalable , open-source LLM written in pure Python/Jax and targeting Google Cloud TPUs and GPUs for training and inference . MaxText achieves high MFUs and scales from single host to very large clusters while staying simple and "optimization-free" thanks to the power of Jax and the XLA compiler.
MaxText... See more
MaxText is a high performance , highly scalable , open-source LLM written in pure Python/Jax and targeting Google Cloud TPUs and GPUs for training and inference . MaxText achieves high MFUs and scales from single host to very large clusters while staying simple and "optimization-free" thanks to the power of Jax and the XLA compiler.
MaxText... See more
google β’ GitHub - google/maxtext: A simple, performant and scalable Jax LLM!
baserun.aiπͺπͺπͺ
Testing & Observability Platform for LLM Apps
From prompt playground to end-to-end tests, baserun helps you ship your LLM apps with confidence and speed.
Testing framework for LLM Part
β‘ LitGPT
Pretrain, finetune, evaluate, and deploy 20+ LLMs on your own data
Uses the latest state-of-the-art techniques:
β flash attention β fp4/8/16/32 β LoRA, QLoRA, Adapter (v1, v2) β FSDP β 1-1000+ GPUs/TPUs
Lightning AI β’ Models β’ Quick start β’ Inference β’ Finetune β’ Pretrain β’ Deploy β’ Features β’ Training recipes (YAML)
Finetune, pretrain and... See more
Pretrain, finetune, evaluate, and deploy 20+ LLMs on your own data
Uses the latest state-of-the-art techniques:
β flash attention β fp4/8/16/32 β LoRA, QLoRA, Adapter (v1, v2) β FSDP β 1-1000+ GPUs/TPUs
Lightning AI β’ Models β’ Quick start β’ Inference β’ Finetune β’ Pretrain β’ Deploy β’ Features β’ Training recipes (YAML)
Finetune, pretrain and... See more