GitHub - roboflow/multimodal-maestro: Effective prompting for Large Multimodal Models like GPT-4 Vision, LLaVA or CogVLM. 🔥
Welcome to prompttools created by Hegel AI! This repo offers a set of open-source, self-hostable tools for experimenting with, testing, and evaluating LLMs, vector databases, and prompts. The core idea is to enable developers to evaluate using familiar interfaces like code, notebooks, and a local playground.... See more
In just a few lines of codes, you can t
Testing framework for LLM Part
Nicolay Gerold added
Open AI released a great Prompt Engineering guide (link in comments).
Here is a summary of their 6 strategies for getting better results when prompting GPT-4 https://t.co/QiBASsbOHD... See more
Abhishek Sivaraman and added
Key unlock: Multimodal models can reason about images, video, or even physical environments without significant tailoring.
Sarah Wang • The Next Token of Progress: 4 Unlocks on the Generative AI Horizon
Darren LI added
LLaVA v1.5, a new open-source multimodal model stepping onto the scene as a contender against GPT-4 with multimodal capabilities. It uses a simple projection matrix to connect the pre-trained CLIP ViT-L/14 vision encoder with Vicuna LLM, resulting in a robust model that can handle images and text. The model is trained in two stages: first, updated ... See more
This AI newsletter is all you need #68
Nicolay Gerold added
The Dawn of LMMs: Preliminary Explorations with GPT-4V(ision)
An analysis of GPT-4V, a large multimodal model with visual understanding, discussing its capabilities, input modes, working modes, prompting techniques, and potential applications in various domains.
browse.arxiv.orgDarren LI added
Hands-on with Gemini: Interacting with multimodal AI
youtu.beYanis Markin added
Arielle Shnaidman and added
promptfoo is a tool for testing and evaluating LLM output quality.... See more
With promptfoo, you can:
Systematically test prompts & models against predefined test cases
Evaluate quality and catch regressions by comparing LLM outputs side-by-side
Speed up evaluations with caching and concurrency
Score outputs automatically by defining test cases
Use as a
Testing framework for LLM Part
Nicolay Gerold added