Agents
π Introduction
ReactAgent is an experimental autonomous agent that uses GPT-4 language model to generate and compose React components from user stories. It is built with React, TailwindCSS, Typescript, Radix UI, Shandcn UI, and OpenAI API.
π Features
ReactAgent is an experimental autonomous agent that uses GPT-4 language model to generate and compose React components from user stories. It is built with React, TailwindCSS, Typescript, Radix UI, Shandcn UI, and OpenAI API.
π Features
- Generate React Components from user stories
- Compose React Components from existing components
- Use a
eylonmiz β’ GitHub - eylonmiz/react-agent: The open-source React.js Autonomous LLM Agent
The Semantic Layer
for every data app
Connect data silos, drive consistent metrics, and power your AI and analytics with context.
for every data app
Connect data silos, drive consistent metrics, and power your AI and analytics with context.
Cube β Semantic Layer for Building Data Applications
Introduction
VT.ai is a multi-modal AI Chatbot Assistant, offering a chat interface to interact with Large Language Models (LLMs) from various providers. Both via remote API or running locally with Ollama.
The application supports multi-modal conversations, seamlessly integrating text, images, and vision processing with LLMs.
[Beta] Multi-modal AI... See more
VT.ai is a multi-modal AI Chatbot Assistant, offering a chat interface to interact with Large Language Models (LLMs) from various providers. Both via remote API or running locally with Ollama.
The application supports multi-modal conversations, seamlessly integrating text, images, and vision processing with LLMs.
[Beta] Multi-modal AI... See more
GitHub - vinhnx/VT.ai: VT.ai - Multi-modal AI Chatbot Assistant
π GPT Researcher
GPT Researcher is an autonomous agent designed for comprehensive online research on a variety of tasks.
The agent can produce detailed, factual and unbiased research reports, with customization options for focusing on relevant resources, outlines, and lessons. Inspired by the recent Plan-and-Solve and RAG (Retrieval Augmented... See more
GPT Researcher is an autonomous agent designed for comprehensive online research on a variety of tasks.
The agent can produce detailed, factual and unbiased research reports, with customization options for focusing on relevant resources, outlines, and lessons. Inspired by the recent Plan-and-Solve and RAG (Retrieval Augmented... See more
assafelovic β’ GitHub - assafelovic/gpt-researcher: GPT based autonomous agent that does online comprehensive research on any given topic
π Introduction
XAgent is an open-source experimental Large Language Model (LLM) driven autonomous agent that can automatically solve various tasks. It is designed to be a general-purpose agent that can be applied to a wide range of tasks. XAgent is still in its early stages, and we are working hard to improve it.
π Our goal is to create a... See more
XAgent is an open-source experimental Large Language Model (LLM) driven autonomous agent that can automatically solve various tasks. It is designed to be a general-purpose agent that can be applied to a wide range of tasks. XAgent is still in its early stages, and we are working hard to improve it.
π Our goal is to create a... See more
OpenBMB β’ GitHub - OpenBMB/XAgent: An Autonomous LLM Agent for Complex Task Solving
In an Agentic UX, an AI agent pulls weight across 3 categories: cognitive, creative, and logistical.
- Cognitive weight β the built-in agent pulls analysis and decision-making weight
- Creative weight β the built-in agent takes on visualization and media creation weight
- Logistical weight β the built-in agent takes tasks on operational and workflow
The agentic era of UX
React.js LLM Agent for next generation codingReactAgent is an experimental autonomous agent that uses GPT-4 language model to generate and compose React components from user stories. It is built with React, TailwindCSS, Typescript, Radix UI, Shandcn UI, and OpenAI API.
ReactAgent - The open-source React.js LLM Agent
Self-Operating Computer Framework
A framework to enable multimodal models to operate a computer.
Using the same inputs and outputs of a human operator, the model views the screen and decides on a series of mouse and keyboard actions to reach an objective.
Key Features
A framework to enable multimodal models to operate a computer.
Using the same inputs and outputs of a human operator, the model views the screen and decides on a series of mouse and keyboard actions to reach an objective.
Key Features
- Compatibility : Designed for various multimodal models.
- Integration : Currently
OthersideAI β’ GitHub - OthersideAI/self-operating-computer: A framework to enable multimodal models to operate a computer.
In the open-source community, there are huge numbers of people leveraging AutoGen in creative ways, and solving surprising problems. One pattern that we see as fundamental is the "generator+critic" pattern, where one agent generates content (writing, code, etc.) and another agent critiques it (finds bugs, etc.) They can iterate until the solution... See more