LLMs
How can we make interacting with conversational models feel more natural?
Every conversational interface to a language model adopts the same pattern:
A chat history sidebar, with each conversation lasting just a few turns
New sessions always begin in a brand-new thread
Every user query must always elicit exactly one response
None of these assumptions... See more
Every conversational interface to a language model adopts the same pattern:
A chat history sidebar, with each conversation lasting just a few turns
New sessions always begin in a brand-new thread
Every user query must always elicit exactly one response
None of these assumptions... See more
Shortwave โ rajhesh.panchanadhan@gmail.com [Gmail alternative]
๐ค Cutting-edge framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
- Why CrewAI
- Getting Started
- Key Features
- Examples
- Local Open Source Models
- CrewAI x AutoGen x ChatDev
- Contribution
- ๐ฌ CrewAI Discord Community
- Hire Consulting
- License
joaomdmoura โข GitHub - joaomdmoura/crewAI: Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
๐ฑ๐ฆ๐ณ๐ง๐ฐ๐ณ๐ฎ๐ข๐ฏ๐ค๐ฆ: it will improve your LLM performance on given use cases (e.g., coding, extracting text, etc.). Mainly, the LLM will specialize in a given task (a specialist will always beat a generalist in its domain)
๐ค๐ฐ๐ฏ๐ต๐ณ๐ฐ๐ญ: you can refine how your model should behave on specific inputs and outputs, resulting in a more robust product
๐ฎ๐ฐ๐ฅ๐ถ๐ญ๐ข๐ณ๐ช๐ป๐ข๐ต๐ช๐ฐ๐ฏ:... See more
๐ค๐ฐ๐ฏ๐ต๐ณ๐ฐ๐ญ: you can refine how your model should behave on specific inputs and outputs, resulting in a more robust product
๐ฎ๐ฐ๐ฅ๐ถ๐ญ๐ข๐ณ๐ช๐ป๐ข๐ต๐ช๐ฐ๐ฏ:... See more
Shortwave โ rajhesh.panchanadhan@gmail.com [Gmail alternative]
Motivation for finetuning
๐บ๐ฒ๐๐ต๐ผ๐ฑ๐ ๐ผ๐ณ ๐ณ๐ถ๐ป๐ฒ-๐๐๐ป๐ถ๐ป๐ด ๐ฎ๐ป ๐ผ๐ฝ๐ฒ๐ป-๐๐ผ๐๐ฟ๐ฐ๐ฒ ๐๐๐ ๐ฒ๐
๐ถ๐t โ
- ๐๐ฐ๐ฏ๐ต๐ช๐ฏ๐ถ๐ฆ๐ฅ ๐ฑ๐ณ๐ฆ-๐ต๐ณ๐ข๐ช๐ฏ๐ช๐ฏ๐จ: utilize domain-specific data to apply the same pre-training process (next token prediction) on the pre-trained (base) model
- ๐๐ฏ๐ด๐ต๐ณ๐ถ๐ค๐ต๐ช๐ฐ๐ฏ ๐ง๐ช๐ฏ๐ฆ-๐ต๐ถ๐ฏ๐ช๐ฏ๐จ: the pre-trained (base) model is fine-tuned on a Q&A dataset to learn to answer questions
- ๐๐ช๐ฏ๐จ๐ญ๐ฆ-๐ต๐ข๐ด๐ฌ ๐ง๐ช๐ฏ๐ฆ-๐ต๐ถ๐ฏ๐ช๐ฏ๐จ: the... See more
- ๐๐ฐ๐ฏ๐ต๐ช๐ฏ๐ถ๐ฆ๐ฅ ๐ฑ๐ณ๐ฆ-๐ต๐ณ๐ข๐ช๐ฏ๐ช๐ฏ๐จ: utilize domain-specific data to apply the same pre-training process (next token prediction) on the pre-trained (base) model
- ๐๐ฏ๐ด๐ต๐ณ๐ถ๐ค๐ต๐ช๐ฐ๐ฏ ๐ง๐ช๐ฏ๐ฆ-๐ต๐ถ๐ฏ๐ช๐ฏ๐จ: the pre-trained (base) model is fine-tuned on a Q&A dataset to learn to answer questions
- ๐๐ช๐ฏ๐จ๐ญ๐ฆ-๐ต๐ข๐ด๐ฌ ๐ง๐ช๐ฏ๐ฆ-๐ต๐ถ๐ฏ๐ช๐ฏ๐จ: the... See more
Shortwave โ rajhesh.panchanadhan@gmail.com [Gmail alternative]
To do this, we employ a technique known as AI-assisted evaluation, alongside traditional metrics for measuring performance. This helps us pick the prompts that lead to better quality outputs, making the end product more appealing to users. AI-assisted evaluation uses best-in-class LLMs (like GPT-4) to automatically critique how well the AI's... See more
Developing Rapidly with Generative AI
A solution is to self-host an open-sourced or custom fine-tuned LLM. Opting for a self-hosted model can reduce costs dramatically - but with additional development time, maintenance overhead, and possible performance implications. Considering self-hosted solutions requires weighing these different trade-offs carefully.
Developing Rapidly with Generative AI
memary: Open-Source Longterm Memory for Autonomous Agents
memary demo
Why use memary?
Agents use LLMs that are currently constrained to finite context windows. memary overcomes this limitation by allowing your agents to store a large corpus of information in knowledge graphs, infer user knowledge through our memory modules, and only retrieve... See more
memary demo
Why use memary?
Agents use LLMs that are currently constrained to finite context windows. memary overcomes this limitation by allowing your agents to store a large corpus of information in knowledge graphs, infer user knowledge through our memory modules, and only retrieve... See more
GitHub - kingjulio8238/memary: Longterm Memory for Autonomous Agents.
Data
Matei Zaharia, Omar Khattab, Lingjiao Chen, et al. โข The Shift From Models to Compound AI Systems
core components of Deep RL that enabled success like AlphaGo: self-play and look-ahead planning.
Self-play is the idea that an agent can improve its gameplay by playing against slightly different versions of itself because itโll progressively encounter more challenging situations. In the space of LLMs, it is almost certain that the largest portion... See more
Self-play is the idea that an agent can improve its gameplay by playing against slightly different versions of itself because itโll progressively encounter more challenging situations. In the space of LLMs, it is almost certain that the largest portion... See more
Shortwave โ rajhesh.panchanadhan@gmail.com [Gmail alternative]
These two components might be some of the most important ideas to improve all of AI.