• from Tips for probabilistic software - jxnl.co by Jason Liu

    Nicolay Gerold added 5d ago

  • Large variety of ready-to-use LLM evaluation metrics (all with explanations) powered by

    ANY

    LLM of your choice, statistical methods, or NLP models that runs

    locally on your machine

    :

    • G-Eval

    • Summarization

    • Answer Relevancy

    • Faithfulness

    • Contextual Recall

    • Contextual Precision

    • RAGAS

    • Hallucination

    • Toxicity

    • Bias

    • etc.

from GitHub - confident-ai/deepeval: The LLM Evaluation Framework

Nicolay Gerold added 17d ago

  • from Beyond customization: build tools that grow with us | thesephist.com

    Nicolay Gerold added 21d ago

  • from Legal documents are pushing text interfaces forward | thesephist.com

    Nicolay Gerold added 21d ago

  • from Navigate, don't search | thesephist.com by Linus Lee

    Nicolay Gerold added 21d ago

  • from GitHub - sqrkl/lm-evaluation-harness: A framework for few-shot evaluation of language models.

    Nicolay Gerold added 23d ago

  • from Shortwave — rajhesh.panchanadhan@gmail.com [Gmail alternative]

    Nicolay Gerold added 25d ago

  • from Shortwave — rajhesh.panchanadhan@gmail.com [Gmail alternative]

    Nicolay Gerold added 1mo ago

  • from Shortwave — rajhesh.panchanadhan@gmail.com [Gmail alternative]

    Nicolay Gerold added 1mo ago