Together AI has announced Mamba-3, a state-space model (SSM) focused on optimizing inference. The news emerged via a post on Together AI's official blog and generated discussions on the Reddit platform.

State-Space Models

State-space models represent an alternative to the more common transformer models, especially in scenarios that require efficient processing of long sequences. Mamba-3 aims to improve performance in this area, potentially offering advantages in terms of speed and resource consumption.

For those evaluating on-premise deployments, there are trade-offs to consider when choosing between transformer architectures and state-space models. AI-RADAR offers analytical frameworks on /llm-onpremise to evaluate these aspects.