AI21 Labs’ new AI model can handle more context than most
the AI industry is moving toward generative AI models with longer contexts
Contexts, or context windows, refer to input data that a model considers before generating output .
AI21 Labs’ Jamba, a new text-generating and -analyzing model, can perform many of the same tasks that models like OpenAI’s ChatGPT
Jamba can handle up to 140,000 tokens while running on a single GPU with at least 80GB of memory
Meta’s Llama 2, by comparison, has a 32,000-token context window — on the smaller side by today’s standards
But what makes Jamba unique is what’s under the hood. It uses a combination of two model architectures
Transformers are the architecture of choice for complex reasoning tasks, powering models like GPT-4 and Google’s Gemini
Jamba in fact uses Mamba as part of the core model — and Dagan claims it delivers three times the throughput on long contexts
while Jamba has been released under the Apache 2.0 license
The model doesn’t have safeguards to prevent it from generating toxic text or mitigations to address potential bias;
See all Latest News from here