Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
Devstral is a collaborative effort between Mistral AI and All Hands AI, resulting in an open-source large language model specifically tailored for software engineering. This model demonstrates remarkable proficiency in navigating intricate codebases, managing edits across numerous files, and addressing practical problems, achieving a notable score of 46.8% on the SWE-Bench Verified benchmark, which is superior to all other open-source models. Based on Mistral-Small-3.1, Devstral boasts an extensive context window supporting up to 128,000 tokens. It is designed for optimal performance on high-performance hardware setups, such as Macs equipped with 32GB of RAM or Nvidia RTX 4090 GPUs, and supports various inference frameworks including vLLM, Transformers, and Ollama. Released under the Apache 2.0 license, Devstral is freely accessible on platforms like Hugging Face, Ollama, Kaggle, Unsloth, and LM Studio, allowing developers to integrate its capabilities into their projects seamlessly. This model not only enhances productivity for software engineers but also serves as a valuable resource for anyone working with code.
Description
Mistral AI has launched two cutting-edge models designed for on-device computing and edge applications, referred to as "les Ministraux": Ministral 3B and Ministral 8B. These innovative models redefine the standards of knowledge, commonsense reasoning, function-calling, and efficiency within the sub-10B category. They are versatile enough to be utilized or customized for a wide range of applications, including managing complex workflows and developing specialized task-focused workers. Capable of handling up to 128k context length (with the current version supporting 32k on vLLM), Ministral 8B also incorporates a unique interleaved sliding-window attention mechanism to enhance both speed and memory efficiency during inference. Designed for low-latency and compute-efficient solutions, these models excel in scenarios such as offline translation, smart assistants that don't rely on internet connectivity, local data analysis, and autonomous robotics. Moreover, when paired with larger language models like Mistral Large, les Ministraux can effectively function as streamlined intermediaries, facilitating function-calling within intricate multi-step workflows, thereby expanding their applicability across various domains. This combination not only enhances performance but also broadens the scope of what can be achieved with AI in edge computing.
API Access
Has API
API Access
Has API
Integrations
AlphaCorp
Arize Phoenix
Continue
Deep Infra
Diaflow
Echo AI
EvalsOne
HumanLayer
Kaggle
Mathstral
Integrations
AlphaCorp
Arize Phoenix
Continue
Deep Infra
Diaflow
Echo AI
EvalsOne
HumanLayer
Kaggle
Mathstral
Pricing Details
$0.1 per million input tokens
Free Trial
Free Version
Pricing Details
Free
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
Mistral AI
Founded
2023
Country
France
Website
mistral.ai/news/devstral
Vendor Details
Company Name
Mistral AI
Founded
2023
Country
France
Website
mistral.ai/news/ministraux/