LLMs
TorchMultimodal (Beta Release)
Introduction
TorchMultimodal is a PyTorch library for training state-of-the-art multimodal multi-task models at scale. It provides:
Introduction
TorchMultimodal is a PyTorch library for training state-of-the-art multimodal multi-task models at scale. It provides:
- A repository of modular and composable building blocks (models, fusion layers, loss functions, datasets and utilities).
- A repository of examples that show how to combine these building
facebookresearch โข GitHub - facebookresearch/multimodal at a33a8b888a542a4578b16972aecd072eff02c1a6
๐บ๐ฒ๐๐ต๐ผ๐ฑ๐ ๐ผ๐ณ ๐ณ๐ถ๐ป๐ฒ-๐๐๐ป๐ถ๐ป๐ด ๐ฎ๐ป ๐ผ๐ฝ๐ฒ๐ป-๐๐ผ๐๐ฟ๐ฐ๐ฒ ๐๐๐ ๐ฒ๐
๐ถ๐t โ
- ๐๐ฐ๐ฏ๐ต๐ช๐ฏ๐ถ๐ฆ๐ฅ ๐ฑ๐ณ๐ฆ-๐ต๐ณ๐ข๐ช๐ฏ๐ช๐ฏ๐จ: utilize domain-specific data to apply the same pre-training process (next token prediction) on the pre-trained (base) model
- ๐๐ฏ๐ด๐ต๐ณ๐ถ๐ค๐ต๐ช๐ฐ๐ฏ ๐ง๐ช๐ฏ๐ฆ-๐ต๐ถ๐ฏ๐ช๐ฏ๐จ: the pre-trained (base) model is fine-tuned on a Q&A dataset to learn to answer questions
- ๐๐ช๐ฏ๐จ๐ญ๐ฆ-๐ต๐ข๐ด๐ฌ ๐ง๐ช๐ฏ๐ฆ-๐ต๐ถ๐ฏ๐ช๐ฏ๐จ: the... See more
- ๐๐ฐ๐ฏ๐ต๐ช๐ฏ๐ถ๐ฆ๐ฅ ๐ฑ๐ณ๐ฆ-๐ต๐ณ๐ข๐ช๐ฏ๐ช๐ฏ๐จ: utilize domain-specific data to apply the same pre-training process (next token prediction) on the pre-trained (base) model
- ๐๐ฏ๐ด๐ต๐ณ๐ถ๐ค๐ต๐ช๐ฐ๐ฏ ๐ง๐ช๐ฏ๐ฆ-๐ต๐ถ๐ฏ๐ช๐ฏ๐จ: the pre-trained (base) model is fine-tuned on a Q&A dataset to learn to answer questions
- ๐๐ช๐ฏ๐จ๐ญ๐ฆ-๐ต๐ข๐ด๐ฌ ๐ง๐ช๐ฏ๐ฆ-๐ต๐ถ๐ฏ๐ช๐ฏ๐จ: the... See more
Shortwave โ rajhesh.panchanadhan@gmail.com [Gmail alternative]
a couple of the top of my head:
- LLM in the loop with preference optimization
- synthetic data generation
- cross modality "distillation" / dictionary remapping
- constrained decoding
r/MachineLearning - Reddit
Additional LLM paradigms beyond RAG
OpenAI is treating its new marketplace seriously now: The brand new GPT store will come with REVENUE SHARING.... (missing in the Plugins launch)
and launching a Stateful Assistants API:
- Persistent Threads (/api/openai/threads)
- Built in Retrieval (chunking etc done for you)
- Code Interpreter (RIP Adv Data Analysis?)
- Speech to Text and Text to... See more
and launching a Stateful Assistants API:
- Persistent Threads (/api/openai/threads)
- Built in Retrieval (chunking etc done for you)
- Code Interpreter (RIP Adv Data Analysis?)
- Speech to Text and Text to... See more
swyx โข Tweet
Menlo Ventures released a report on โThe State of Generative AI in the Enterpriseโ and found that adoption is trailing the hype. Details below:
Generative AI still represents less than 1% of cloud spend by surveyed enterprises, including just an 8% increase in 2023.
Safety and ROI continue to be prime concerns, and the tangible advantages of being... See more
Generative AI still represents less than 1% of cloud spend by surveyed enterprises, including just an 8% increase in 2023.
Safety and ROI continue to be prime concerns, and the tangible advantages of being... See more
Shortwave โ rajhesh.panchanadhan@gmail.com [Gmail alternative]
Today, weโre releasing the Assistants API, our first step towards helping developers build agent-like experiences within their own applications. An assistant is a purpose-built AI that has specific instructions, leverages extra knowledge, and can call models and tools to perform tasks. The new Assistants API provides new capabilities such as Code... See more
New models and developer products announced at DevDay
First of all, I'd say you have a bigger problem where your company is trying to find nails with a hammer. That is where your sentiment comes from, and could be an obstacle for both you and the company. It's the same deal when I see people keep on talking about RAG, and nowadays "modular RAG", when really, you could treat everything as a software... See more
r/MachineLearning - Reddit
- Multiple indices. Splitting the document corpus up into multiple indices and then routing queries based on some criteria. This means that the search is over a much smaller set of documents rather than the entire dataset. Again, it is not always useful, but it can be helpful for certain datasets. The same approach works with the LLMs themselves.
Matt Rickard โข Improving RAG: Strategies
The way that most RLHF is done to date has the entire response from a language model get an associated score. To anyone with an RL background, this is disappointing, because it limits the ability for RL methods to make connections about the value of each sub-component of text. Futures have been pointed to where this multi-step optimization comes at... See more