GitHub - vinhnx/VT.ai: VT.ai - Multi-modal AI Chatbot Assistant
TypingMind — The best chat UI for AI models (ChatGPT, Gemini, Claude, and more)
typingmind.comSteven and added
Create OpenAI like AI assistant with Llama-3 deployed locally on your computer (100% free and without internet): https://t.co/mLIOGM6Ly1
Nathan Storey added
https://github.com/huggingface/chat-ui - Amazing clean UI with very good web search, my go to currently. (they added the ability to do it all locally very recently!)
https://github.com/oobabooga/text-generation-webui - Best overall, supports any model format and has many extensions
https://github.com/ParisNeo/lollms-webui/ - Has PDF, stable diffusion... See more
https://github.com/oobabooga/text-generation-webui - Best overall, supports any model format and has many extensions
https://github.com/ParisNeo/lollms-webui/ - Has PDF, stable diffusion... See more
r/LocalLLaMA - Reddit
Nicolay Gerold added
Welcome to prompttools created by Hegel AI! This repo offers a set of open-source, self-hostable tools for experimenting with, testing, and evaluating LLMs, vector databases, and prompts. The core idea is to enable developers to evaluate using familiar interfaces like code, notebooks, and a local playground.... See more
In just a few lines of codes, you can t
Testing framework for LLM Part
Nicolay Gerold added
Hands-on with Gemini: Interacting with multimodal AI
youtu.beYanis Markin added
OpenGPTs
This is an open source effort to create a similar experience to OpenAI's GPTs. It builds upon LangChain, LangServe and LangSmith. OpenGPTs gives you more control, allowing you to configure:
This is an open source effort to create a similar experience to OpenAI's GPTs. It builds upon LangChain, LangServe and LangSmith. OpenGPTs gives you more control, allowing you to configure:
- The LLM you use (choose between the 60+ that LangChain offers)
- The prompts you use (use LangSmith to debug those)
- The tools you give it (choose from LangC
langchain-ai • langchain-ai/opengpts
Nicolay Gerold added
LLaVA v1.5, a new open-source multimodal model stepping onto the scene as a contender against GPT-4 with multimodal capabilities. It uses a simple projection matrix to connect the pre-trained CLIP ViT-L/14 vision encoder with Vicuna LLM, resulting in a robust model that can handle images and text. The model is trained in two stages: first, updated ... See more
This AI newsletter is all you need #68
Nicolay Gerold added