Sunday, June 8, 2025
News PouroverAI
Visit PourOver.AI
No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing
News PouroverAI
No Result
View All Result

AI21 Labs Breaks New Ground with ‘Jamba’: The Pioneering Hybrid SSM-Transformer Large Language Model

March 28, 2024
in Data Science & ML
Reading Time: 3 mins read
0 0
A A
0
Share on FacebookShare on Twitter


In an era where the demand for smarter, faster, and more efficient artificial intelligence (AI) solutions is continuously on the rise, AI21 Labs’ unveiling of Jamba marks a significant leap forward. Jamba, a pioneering SSM-Transformer model, heralds a new chapter in AI technology by melding the Mamba Structured State Space model (SSM) with the proven efficiency of the traditional Transformer architecture, setting a new benchmark for performance and efficiency in large language models (LLMs).

The Innovation Behind Jamba

At the heart of Jamba lies an integrated blend of Mamba and Transformer architectures designed to address the inherent limitations of each system while leveraging their strengths. Unlike conventional models predominantly based on the Transformer architecture—such as GPT, Gemini, and Llama—Jamba introduces a hybrid approach. It features a remarkable context window of 256K tokens, equivalent to around 210 pages of text, and can fit up to 140K tokens on a single 80GB GPU. This capability significantly surpasses the current standards, like Meta’s Llama 2, which manages a 32,000-token context window.

Jamba’s hybrid architecture combines Transformer, Mamba, and mixture-of-experts (MoE) layers, optimizing memory, throughput, and performance. The model operates on a principle that utilizes mixture-of-experts layers to draw on just 12B of its available 52B parameters during inference, allowing for increased efficiency without sacrificing the model’s power or speed.

Unprecedented Throughput and Efficiency

One of the most significant advantages of Jamba is its ability to deliver three times the throughput on long contexts when compared to Transformer-based models of a similar size, such as Mixtral 8x7B. This efficiency is made possible through its unique architectural composition, which includes a mix of attention, Mamba, and MoE layers. This structure not only enhances the model’s performance but also ensures high throughput and memory optimization.

Moreover, Jamba’s architecture follows a blocks-and-layers approach, which incorporates an attention or Mamba layer followed by a multi-layer perceptron (MLP), achieving an optimal ratio that maximizes quality and throughput on a single GPU. This approach allows for the accommodation of common inference workloads without memory constraints.

Open Access and Future Prospects

AI21 Labs has released Jamba with open weights under the Apache 2.0 license, making it available on Hugging Face and soon on the NVIDIA API catalog as an NVIDIA NIM inference microservice. This move not only democratizes access to Jamba’s advanced capabilities but also invites the AI community to explore, refine, and build upon this innovative architecture.

Although currently released as a research model without the necessary safeguards for commercial use, AI21 Labs plans to unveil a fine-tuned, safer version in the coming weeks. This progression underscores the industry’s commitment to enhancing AI’s performance, efficiency, and accessibility, paving the way for the next generation of AI models.

Key Takeaways

Jamba is the first production-grade AI model that combines the Mamba Structured State Space model (SSM) with the Transformer architecture, addressing the limitations of each while harnessing their strengths.

With a context window of 256K tokens and the ability to fit 140K tokens on a single 80GB GPU, Jamba significantly outperforms existing models in terms of memory efficiency and context handling.

It delivers three times the throughput on long contexts compared to similar-sized Transformer-based models, marking a new efficiency benchmark.

Jamba has been released with open weights under the Apache 2.0 license, available on Hugging Face and soon on the NVIDIA API catalog, fostering community engagement and further innovation.

The release of a commercial-ready version of Jamba is anticipated, which will likely set new standards for AI model performance and application.

Jamba’s introduction by AI21 Labs not only represents a technical milestone but also a shift towards more accessible, efficient, and powerful AI models. As the AI community continues to evolve, the principles and innovations behind Jamba will undoubtedly influence future developments in AI technology.



Source link

Tags: AI21breaksgroundHybridJambaLabslanguageLargemodelPioneeringSSMTransformer
Previous Post

Probabilistic time series forecasting with compositional bayesian neural networks – Google Research Blog

Next Post

Bitwise Seeks Approval for Ethereum ETF Amid SEC Uncertainty

Related Posts

AI Compared: Which Assistant Is the Best?
Data Science & ML

AI Compared: Which Assistant Is the Best?

June 10, 2024
5 Machine Learning Models Explained in 5 Minutes
Data Science & ML

5 Machine Learning Models Explained in 5 Minutes

June 7, 2024
Cohere Picks Enterprise AI Needs Over ‘Abstract Concepts Like AGI’
Data Science & ML

Cohere Picks Enterprise AI Needs Over ‘Abstract Concepts Like AGI’

June 7, 2024
How to Learn Data Analytics – Dataquest
Data Science & ML

How to Learn Data Analytics – Dataquest

June 6, 2024
Adobe Terms Of Service Update Privacy Concerns
Data Science & ML

Adobe Terms Of Service Update Privacy Concerns

June 6, 2024
Build RAG applications using Jina Embeddings v2 on Amazon SageMaker JumpStart
Data Science & ML

Build RAG applications using Jina Embeddings v2 on Amazon SageMaker JumpStart

June 6, 2024
Next Post
Bitwise Seeks Approval for Ethereum ETF Amid SEC Uncertainty

Bitwise Seeks Approval for Ethereum ETF Amid SEC Uncertainty

Gamida Cell to delist as Highbridge takes full ownership

Gamida Cell to delist as Highbridge takes full ownership

World’s biggest shipping firm dumps port cargo problem on US companies

World's biggest shipping firm dumps port cargo problem on US companies

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

  • Trending
  • Comments
  • Latest
23 Plagiarism Facts and Statistics to Analyze Latest Trends

23 Plagiarism Facts and Statistics to Analyze Latest Trends

June 4, 2024
Accenture creates a regulatory document authoring solution using AWS generative AI services

Accenture creates a regulatory document authoring solution using AWS generative AI services

February 6, 2024
Managing PDFs in Node.js with pdf-lib

Managing PDFs in Node.js with pdf-lib

November 16, 2023
Graph neural networks in TensorFlow – Google Research Blog

Graph neural networks in TensorFlow – Google Research Blog

February 6, 2024
13 Best Books, Courses and Communities for Learning React — SitePoint

13 Best Books, Courses and Communities for Learning React — SitePoint

February 4, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

Can You Guess What Percentage Of Their Wealth The Rich Keep In Cash?

June 10, 2024
AI Compared: Which Assistant Is the Best?

AI Compared: Which Assistant Is the Best?

June 10, 2024
How insurance companies can use synthetic data to fight bias

How insurance companies can use synthetic data to fight bias

June 10, 2024
5 SLA metrics you should be monitoring

5 SLA metrics you should be monitoring

June 10, 2024
From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

From Low-Level to High-Level Tasks: Scaling Fine-Tuning with the ANDROIDCONTROL Dataset

June 10, 2024
UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

UGRO Capital: Targeting to hit milestone of Rs 20,000 cr loan book in 8-10 quarters: Shachindra Nath

June 10, 2024
Facebook Twitter LinkedIn Pinterest RSS
News PouroverAI

The latest news and updates about the AI Technology and Latest Tech Updates around the world... PouroverAI keeps you in the loop.

CATEGORIES

  • AI Technology
  • Automation
  • Blockchain
  • Business
  • Cloud & Programming
  • Data Science & ML
  • Digital Marketing
  • Front-Tech
  • Uncategorized

SITEMAP

  • Disclaimer
  • Privacy Policy
  • DMCA
  • Cookie Privacy Policy
  • Terms and Conditions
  • Contact us

Copyright © 2023 PouroverAI News.
PouroverAI News

No Result
View All Result
  • Home
  • AI Tech
  • Business
  • Blockchain
  • Data Science & ML
  • Cloud & Programming
  • Automation
  • Front-Tech
  • Marketing

Copyright © 2023 PouroverAI News.
PouroverAI News

Welcome Back!

Login to your account below

Forgotten Password? Sign Up

Create New Account!

Fill the forms bellow to register

All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In