Models
4. Introducing Stable LM 3B: Bringing Sustainable, High-Performance Language Models to Smart Devices
Stability AI introduced Stable LM 3B, a high-performing language model designed for smart devices. With 3 billion parameters, it outperforms state-of-the-art 3B models and reduces operating costs and power consumption. The model enables a broader... See more
Stability AI introduced Stable LM 3B, a high-performing language model designed for smart devices. With 3 billion parameters, it outperforms state-of-the-art 3B models and reduces operating costs and power consumption. The model enables a broader... See more
This AI newsletter is all you need #68
We are excited to release the first version of our multimodal assistant Yasa-1, a language assistant with visual and auditory sensors that can take actions via code execution.
We trained Yasa-1 from scratch, including pretraining base models from ground zero, aligning them, as well as heavily optimizing both our training and serving infrastructure.... See more
We trained Yasa-1 from scratch, including pretraining base models from ground zero, aligning them, as well as heavily optimizing both our training and serving infrastructure.... See more
Announcing our Multimodal AI Assistant - Reka AI
The Nemotron-3 8B family is available in the Azure AI Model Catalog, HuggingFace, and the NVIDIA AI Foundation Model hub on the NVIDIA NGC Catalog. It includes base, chat, and question-and-answer (Q&A) models that are designed to solve a variety of downstream tasks. Table 1 shows the full family of foundation models.
Model
Variant
Key Benefit
Base
N... See more
Model
Variant
Key Benefit
Base
N... See more
NVIDIA Technical Blog | News and tutorials for developers, data ...
Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr-7B-α is the first model in the series, and is a fine-tuned version of mistralai/Mistral-7B-v0.1 that was trained on on a mix of publicly available, synthetic datasets using Direct Preference Optimization (DPO). We found that removing the in-built alignment... See more
HuggingFaceH4/zephyr-7b-alpha · Hugging Face
Macaw-LLM: Multi-Modal Language Modeling with Image, Audio, Video, and Text Integration
1 2 Chenyang Lyu, 3 Minghao Wu, 1 * Longyue Wang, 1 Xinting Huang,
1 Bingshuai Liu, 1 Zefeng Du, 1 Shuming Shi, 1 Zhaopeng Tu
1 Tencent AI Lab, 2 Dublin City University, 3 Monash University
* Longyue Wang is the corresponding author: vinnlywang@tencent.com
Macaw... See more
1 2 Chenyang Lyu, 3 Minghao Wu, 1 * Longyue Wang, 1 Xinting Huang,
1 Bingshuai Liu, 1 Zefeng Du, 1 Shuming Shi, 1 Zhaopeng Tu
1 Tencent AI Lab, 2 Dublin City University, 3 Monash University
* Longyue Wang is the corresponding author: vinnlywang@tencent.com
Macaw... See more
lyuchenyang • GitHub - lyuchenyang/Macaw-LLM: Macaw-LLM: Multi-Modal Language Modeling with Image, Video, Audio, and Text Integration
- Microsoft introduced
FOD#27: "Now And Then"
ScaleCrafter is capable of generating images with resolution of 4096 x 4096 and results with resolution of 2048 x 1152 based on pre-trained diffusion models on a lower resolution. Notably, our approach needs no extra training/optimziation .
YingqingHe • GitHub - YingqingHe/ScaleCrafter: Official implementation of ScaleCrafter for higher-resolution visual generation at inference time.
Glaive-coder-7b
Glaive-coder-7b is a 7B parameter code model trained on a dataset of ~140k programming related problems and solutions generated from Glaive’s synthetic data generation platform.
The model is fine-tuned on the CodeLlama-7b model.
Usage:
The model is trained to act as a code assistant, and can do both single instruction following and... See more
Glaive-coder-7b is a 7B parameter code model trained on a dataset of ~140k programming related problems and solutions generated from Glaive’s synthetic data generation platform.
The model is fine-tuned on the CodeLlama-7b model.
Usage:
The model is trained to act as a code assistant, and can do both single instruction following and... See more
glaiveai/glaive-coder-7b · Hugging Face
Supported Models
Suggest Edits
Where possible, we try to match the Hugging Face implementation. We are open to adjusting the API, so please reach out with feedback regarding these details.
Model
Context Length
Model Type
codellama-34b-instruct
16384
Chat Completion
llama-2-70b-chat
4096
Chat Completion
mistral-7b-instruct
4096 [1]
Chat Completion
pplx-7b-c... See more
Suggest Edits
Where possible, we try to match the Hugging Face implementation. We are open to adjusting the API, so please reach out with feedback regarding these details.
Model
Context Length
Model Type
codellama-34b-instruct
16384
Chat Completion
llama-2-70b-chat
4096
Chat Completion
mistral-7b-instruct
4096 [1]
Chat Completion
pplx-7b-c... See more
Supported Models
Models by perplexitiy, among other their online model with access to the internet.