GitHub - microsoft/LLMLingua: To speed up LLMs' inference and enhance LLM's perceive of key information, compress the prompt and KV-Cache, which achieves up to 20x compression with minimal performance loss.
updated 8mo ago
updated 8mo ago
Nicolay Gerold added
Easy data labelling for NLP tasks.
Nicolay Gerold added
Nicolay Gerold added