Early this year, we trained a 70B model optimized for reasoning and coding. This model roughly matches LLAMA 3 70B despite being trained on 7x less data.
Today, we’re releasing a toolkit to help others do the same, including:
• 11 sanitized and extended NLP reasoning benchmarks including ARC... See more