A new leader in the small models category by Mistral AI, with image understanding capabilities and an extended 128k context length. Apache 2.0 license.
Use our main calculator for more detailed estimates including input/output combinations.
Key Features
Image understanding capabilities
128K token context window
Open source (Apache 2.0)
Efficient and performant for its size
Common Use Cases
Multimodal applications (text and image)
Cost-effective general tasks
Applications requiring longer context
Frequently Asked Questions
What are Mistral AI's key model offerings in late 2024?
Mistral AI offers a range of models, including open-weight models like Mistral 7B and Mixtral 8x7B (a Mixture of Experts model), specialized models like Codestral for coding, and commercial API models such as Mistral Small, Mistral Large (and potentially a newer "Mistral Large 2" or equivalent). They focus on high performance, efficiency, and often strong multilingual capabilities.
What is Codestral, and what are its main strengths?
Codestral is an open-weight generative model from Mistral AI specifically designed for code-related tasks. It supports a wide range of programming languages (80+) and excels at code generation, completion, explanation, and debugging. It typically has a large context window suitable for understanding code repositories.
How does Mistral AI differentiate its commercial models (e.g., Mistral Large 2) from its open-weight models?
Mistral AI's commercial models, accessed via their API, generally offer higher performance, larger context windows, and potentially more advanced features (like optimized function calling or RAG capabilities) compared to their open-weight counterparts. They are designed for enterprise use cases requiring reliability and cutting-edge performance, while their open models promote broader accessibility and innovation.