Skip to content
Thoughts on GitHub Models?

AI21 Jamba 1.5 Mini

A 52B parameters (12B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.
Context
262k input · 4k output
Training date
Undisclosed
Rate limit tier
Provider support
Try AI21 Jamba 1.5 Mini
Azure hosted. AI powered, can make mistakes. . Subject to Product Terms & Privacy Statement. Not intended for production/sensitive data.
What are some of the most famous works of Shakespeare?
Can you explain the concept of time dilation in physics?
What are some common features of Gothic architecture?

Model navigation navigation

AI21 Labs

Jamba 1.5 Mini is a state-of-the-art, hybrid SSM-Transformer instruction following foundation model. It's a Mixture-of-Expert model with 52B total parameters and 12B active parameters. The Jamba family of models are the most powerful & efficient long-context models on the market, offering a 256K context window, the longest available.. For long context input, they deliver up to 2.5X faster inference than leading models of comparable sizes. Jamba supports function calling/tool use, structured output (JSON), and grounded generation with citation mode and documents API. Jamba officially supports English, French, Spanish, Portuguese, German, Arabic and Hebrew, but can also work in many other languages.

Model Developer Name: AI21 Labs

Model Architecture

Jamba 1.5 Mini is a state-of-the-art, hybrid SSM-Transformer instruction following foundation model

Model Variations

52B total parameters and 12B active parameters

Model Input

Model inputs text only.

Model Output

Model generates text only.

Model Dates

Jamba 1.5 Mini was trained in Q3 2024 with data covering through early March 2024.

​​​Model Information Table

Name Params Content Length
Jamba 1.5 Mini 52B (12B active) 256K
Jamba 1.5 Large 398B (94B active) 256K

Languages

 (7)
English, French, Spanish, Portuguese, German, Arabic, and Hebrew

About

A 52B parameters (12B active) multilingual model, offering a 256K long context window, function calling, structured output, and grounded generation.
Context
262k input · 4k output
Training date
Undisclosed
Rate limit tier
Provider support

Languages

 (7)
English, French, Spanish, Portuguese, German, Arabic, and Hebrew