Overview

Built on the novel Mamba-Transformer architecture, these highly efficient and powerful models push the boundaries of AI. They deliver unmatched speed and quality and feature the longest context window (256K tokens) among open models.

Jamba Large

Our most powerful and advanced model, designed to handle complex tasks at enterprise scale with superior performance.

Jamba Mini

The most efficient model in its size class, optimized for speed and quality with a smaller footprint.

Key Benefits

Privately deployable without compromising on quality

Retain total control over your data, with zero data visibility for the model vendor by deploying Jamba privately in your VPC or on-premises. Ideal for organizations or use cases handling regulated data (e.g., in finance or healthcare) or confidential proprietary data. With comparable quality to market-leading models, you get superior quality with total data privacy.

Open models

With the weights available to be downloaded into your environment, Jamba can be customized to your domain and use case. This is a model that’s fully yours.

Long context performance

With a 256K context window, Jamba excels on the kinds of use cases enterprises rely on in their workflows, from analyzing lengthy documents to enhancing RAG workflows at the retrieval stage. And due to its efficiency-optimized hybrid architecture, it can do all this at a lower cost than competitors.

Self Deployment Options

Jamba models are ideal for enterprises that need to maintain full control over data and performance. Deployment options include:

  • Cloud-hosted VPC – Isolate and scale your deployment in a virtual private cloud environment.
  • On-premises environments – Run Jamba entirely within your own infrastructure to meet strict compliance or latency requirements.
  • Custom hybrid solutions – Combine cloud and on-premises capabilities to fit your architecture, performance, and governance needs.

Supported Languages

Jamba models officially support 9 languages:

  • English, Spanish, French, Portuguese, Italian, Dutch, German, Arabic, Hebrew.

Model Details

ModelModel SizeMax TokensVersionSnapshotAPI Endpoint
Jamba Large398B parameters
(94B active)
256K1.72025-07jamba-large
Jamba Mini52B parameters
(12B active)
256K1.72025-07jamba-mini

Engineers and data scientists at AI21 labs created the model to help developers and businesses leverage AI to build real-world products with tangible value. Jamba Mini and Jamba Large support zero-shot instruction-following and multi-language support. The Jamba models also provide developers with industry-leading APIs that perform a wide range of productivity tasks designed for commercial use.

  • Organization developing model: AI21 Labs
  • Model date: July 3rd, 2025
  • Model type: Joint Attention and Mamba (Jamba)
  • Knowledge cutoff date August 22nd, 2024
  • Input Modality: Text
  • Output Modality: Text
  • License: Jamba open model license
  • Contact: info@ai21.com

API Versioning

We advise using dated versions of the Jamba API to avoid disruptions from model updates and breaking changes. Also, please note that some endpoints will be deprecated soon.

Here are the details of the available versions:

  • jamba-large currently points to jamba-large-1.7-2025-07
  • jamba-mini currently points to jamba-mini-1.7-2025-07
  • jamba-large-1.7 points to jamba-large-1.7-2025-07
  • jamba-mini-1.7 points to jamba-mini-1.7-2025-07

Model Deprecation

ModelSnapshotAPI EndpointDeprecation Date
Jamba Large 1.62025-03jamba-large-1.6-2025-032025-03-08
Jamba Mini 1.62025-03jamba-mini-1.6-2025-032025-03-08
Jamba Large 1.52024-08jamba-large-1.5-2024-082025-05-06
Jamba Mini 1.52024-08jamba-mini-1.5-2024-082025-05-06

Model Compliance and Certifications

Ethical Considerations

AI21 Labs is on a mission to supercharge human productivity with machines working alongside humans as thought partners, thereby promoting human welfare and prosperity. To deliver its promise, this technology must be deployed and used in a responsible and sustainable way, taking into consideration potential risks, including malicious use by bad actors, accidental misuse, and broader societal harms. We take these risks extremely seriously and put measures in place to mitigate them.

AI21 provides open access to Jamba that can be used to power a large variety of useful applications. We believe it is important to ensure that this technology is used in a responsible way, while allowing developers the freedom they need to experiment rapidly and deploy solutions at scale. Overall, we view the safe implementation of this technology as a partnership and collaboration between AI21 and our customers and encourage engagement and dialogue to raise the bar on responsible usage.

In order to use Jamba, you are required to comply with our Terms of Service and with the following Usage Guidelines.

Please check these usage guidelines periodically, as they may be updated from time to time. For any questions, clarifications or concerns, please contact safety@ai21.com.

Limitations

There are a number of limitations inherent to neural networks technology that apply to Jamba. These limitations require explanation and carry important caveats for the application and usage of Jamba.

  • Accuracy: Jamba, like other large pretrained language models, lacks important context about the world because it is trained on textual data and is not grounded in other modalities of experience such as video, real-world physical interaction, and human feedback. Like all language models, Jamba is far more accurate when responding to inputs similar to its training datasets. Novel inputs tend to generate higher variance in its output.
  • Coherence and consistency: Responses from Jamba are sometimes inconsistent, contradictory, or contain seemingly random sentences and paragraphs.
  • Western/English bias: Jamba is trained primarily on English-language text from the internet, and is best suited to classifying, searching, summarizing, and generating English text. Furthermore, Jamba has a tendency to hold and amplify the biases contained in its training dataset. As a result, groups of people who were not involved in the creation of the training data can be underrepresented, and stereotypes and prejudices can be perpetuated. Racial, religious, gender, socioeconomic, and other categorizations of human groups can be considered among these factors.
  • Explainability: It is difficult to explain or predict how Jamba will respond without additional training and fine tuning. This is a common issue with neural networks of this scope and scale.
  • Recency: Jamba was trained on a dataset created in March 2024, and therefore has no knowledge of events that have occurred after that date. We update our models regularly to keep them as current as possible, but there are notable gaps and inaccuracies in responses as a result of this lack of recency.

Ready to take your projects to the next level?
Access Jamba via the AI21 Studio API or deploy privately.
See our Quick Start Guide for integration steps, SDKs, and sample code.

Overview

Built on the novel Mamba-Transformer architecture, these highly efficient and powerful models push the boundaries of AI. They deliver unmatched speed and quality and feature the longest context window (256K tokens) among open models.

Jamba Large

Our most powerful and advanced model, designed to handle complex tasks at enterprise scale with superior performance.

Jamba Mini

The most efficient model in its size class, optimized for speed and quality with a smaller footprint.

Key Benefits

Privately deployable without compromising on quality

Retain total control over your data, with zero data visibility for the model vendor by deploying Jamba privately in your VPC or on-premises. Ideal for organizations or use cases handling regulated data (e.g., in finance or healthcare) or confidential proprietary data. With comparable quality to market-leading models, you get superior quality with total data privacy.

Open models

With the weights available to be downloaded into your environment, Jamba can be customized to your domain and use case. This is a model that’s fully yours.

Long context performance

With a 256K context window, Jamba excels on the kinds of use cases enterprises rely on in their workflows, from analyzing lengthy documents to enhancing RAG workflows at the retrieval stage. And due to its efficiency-optimized hybrid architecture, it can do all this at a lower cost than competitors.

Self Deployment Options

Jamba models are ideal for enterprises that need to maintain full control over data and performance. Deployment options include:

  • Cloud-hosted VPC – Isolate and scale your deployment in a virtual private cloud environment.
  • On-premises environments – Run Jamba entirely within your own infrastructure to meet strict compliance or latency requirements.
  • Custom hybrid solutions – Combine cloud and on-premises capabilities to fit your architecture, performance, and governance needs.

Supported Languages

Jamba models officially support 9 languages:

  • English, Spanish, French, Portuguese, Italian, Dutch, German, Arabic, Hebrew.

Model Details

ModelModel SizeMax TokensVersionSnapshotAPI Endpoint
Jamba Large398B parameters
(94B active)
256K1.72025-07jamba-large
Jamba Mini52B parameters
(12B active)
256K1.72025-07jamba-mini

Engineers and data scientists at AI21 labs created the model to help developers and businesses leverage AI to build real-world products with tangible value. Jamba Mini and Jamba Large support zero-shot instruction-following and multi-language support. The Jamba models also provide developers with industry-leading APIs that perform a wide range of productivity tasks designed for commercial use.

  • Organization developing model: AI21 Labs
  • Model date: July 3rd, 2025
  • Model type: Joint Attention and Mamba (Jamba)
  • Knowledge cutoff date August 22nd, 2024
  • Input Modality: Text
  • Output Modality: Text
  • License: Jamba open model license
  • Contact: info@ai21.com

API Versioning

We advise using dated versions of the Jamba API to avoid disruptions from model updates and breaking changes. Also, please note that some endpoints will be deprecated soon.

Here are the details of the available versions:

  • jamba-large currently points to jamba-large-1.7-2025-07
  • jamba-mini currently points to jamba-mini-1.7-2025-07
  • jamba-large-1.7 points to jamba-large-1.7-2025-07
  • jamba-mini-1.7 points to jamba-mini-1.7-2025-07

Model Deprecation

ModelSnapshotAPI EndpointDeprecation Date
Jamba Large 1.62025-03jamba-large-1.6-2025-032025-03-08
Jamba Mini 1.62025-03jamba-mini-1.6-2025-032025-03-08
Jamba Large 1.52024-08jamba-large-1.5-2024-082025-05-06
Jamba Mini 1.52024-08jamba-mini-1.5-2024-082025-05-06

Model Compliance and Certifications

Ethical Considerations

AI21 Labs is on a mission to supercharge human productivity with machines working alongside humans as thought partners, thereby promoting human welfare and prosperity. To deliver its promise, this technology must be deployed and used in a responsible and sustainable way, taking into consideration potential risks, including malicious use by bad actors, accidental misuse, and broader societal harms. We take these risks extremely seriously and put measures in place to mitigate them.

AI21 provides open access to Jamba that can be used to power a large variety of useful applications. We believe it is important to ensure that this technology is used in a responsible way, while allowing developers the freedom they need to experiment rapidly and deploy solutions at scale. Overall, we view the safe implementation of this technology as a partnership and collaboration between AI21 and our customers and encourage engagement and dialogue to raise the bar on responsible usage.

In order to use Jamba, you are required to comply with our Terms of Service and with the following Usage Guidelines.

Please check these usage guidelines periodically, as they may be updated from time to time. For any questions, clarifications or concerns, please contact safety@ai21.com.

Limitations

There are a number of limitations inherent to neural networks technology that apply to Jamba. These limitations require explanation and carry important caveats for the application and usage of Jamba.

  • Accuracy: Jamba, like other large pretrained language models, lacks important context about the world because it is trained on textual data and is not grounded in other modalities of experience such as video, real-world physical interaction, and human feedback. Like all language models, Jamba is far more accurate when responding to inputs similar to its training datasets. Novel inputs tend to generate higher variance in its output.
  • Coherence and consistency: Responses from Jamba are sometimes inconsistent, contradictory, or contain seemingly random sentences and paragraphs.
  • Western/English bias: Jamba is trained primarily on English-language text from the internet, and is best suited to classifying, searching, summarizing, and generating English text. Furthermore, Jamba has a tendency to hold and amplify the biases contained in its training dataset. As a result, groups of people who were not involved in the creation of the training data can be underrepresented, and stereotypes and prejudices can be perpetuated. Racial, religious, gender, socioeconomic, and other categorizations of human groups can be considered among these factors.
  • Explainability: It is difficult to explain or predict how Jamba will respond without additional training and fine tuning. This is a common issue with neural networks of this scope and scale.
  • Recency: Jamba was trained on a dataset created in March 2024, and therefore has no knowledge of events that have occurred after that date. We update our models regularly to keep them as current as possible, but there are notable gaps and inaccuracies in responses as a result of this lack of recency.

Ready to take your projects to the next level?
Access Jamba via the AI21 Studio API or deploy privately.
See our Quick Start Guide for integration steps, SDKs, and sample code.