AI21 Labs’ new AI model can handle more context than most

the AI industry is moving toward generative AI models with longer contexts

Contexts, or context windows, refer to input data  that a model considers before generating output .

AI21 Labs’ Jamba, a new text-generating and -analyzing model, can perform many of the same tasks that models like OpenAI’s ChatGPT

Jamba can handle up to 140,000 tokens while running on a single GPU with at least 80GB of memory

Meta’s Llama 2, by comparison, has a 32,000-token context window — on the smaller side by today’s standards

But what makes Jamba unique is what’s under the hood. It uses a combination of two model architectures

Transformers are the architecture of choice for complex reasoning tasks, powering models like GPT-4 and Google’s Gemini

Jamba in fact uses Mamba as part of the core model — and Dagan claims it delivers three times the throughput on long contexts

while Jamba has been released under the Apache 2.0 license

The model doesn’t have safeguards to prevent it from generating toxic text or mitigations to address potential bias;

See all Latest News from here