You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/capabilities/embeddings.mdx
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -13,7 +13,7 @@ Embeddings are vectorial representations of text that capture the semantic meani
13
13
<imgsrc="https://colab.research.google.com/assets/colab-badge.svg"alt="Open In Colab"/>
14
14
</a>
15
15
16
-
## Mistral Embeddings API
16
+
## Mistral Embed API
17
17
To generate text embeddings using Mistral AI's embeddings API, we can make a request to the API endpoint and specify the embedding model `mistral-embed`, along with providing a list of input texts. The API will then return the corresponding embeddings as numerical vectors, which can be used for further analysis or processing in NLP applications.
Copy file name to clipboardExpand all lines: docs/getting-started/Open-weight-models.mdx
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,13 +1,13 @@
1
1
---
2
2
id: open_weight_models
3
-
title: Open-weight models
3
+
title: Apache 2.0 models
4
4
sidebar_position: 1.4
5
5
---
6
6
7
7
We open-source both pre-trained models and instruction-tuned models. These models are not tuned for safety as we want to empower users to test and refine moderation based on their use cases. For safer models, follow our [guardrailing tutorial](/capabilities/guardrailing).
8
8
9
9
## License
10
-
- Mistral 7B, Mixtral 8x7B, Mixtral 8x22B, Codestral Mamba, Mathstral, and Mistral NeMo are under [Apache 2 License](https://choosealicense.com/licenses/apache-2.0/), which permits their use without any constraints.
10
+
- Mistral 7B, Mixtral 8x7B, Mixtral 8x22B, Codestral Mamba, Mathstral, and Mistral Nemo are under [Apache 2 License](https://choosealicense.com/licenses/apache-2.0/), which permits their use without any constraints.
11
11
- Codestral is under [Mistral AI Non-Production (MNPL) License](https://mistral.ai/licences/MNPL-0.1.md).
12
12
- Mistral Large is under [Mistral Research License](https://mistral.ai/licenses/MRL-0.1.md).
13
13
@@ -29,8 +29,8 @@ We open-source both pre-trained models and instruction-tuned models. These model
Copy file name to clipboardExpand all lines: docs/getting-started/changelog.mdx
+1-1Lines changed: 1 addition & 1 deletion
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -11,7 +11,7 @@ July 24, 2024
11
11
- We added fine-tuning support for Codestral, Mistral Nemo and Mistral Large. Now the model choices for fine-tuning are `open-mistral-7b` (v0.3), `mistral-small-latest` (`mistral-small-2402`), `codestral-latest` (`codestral-2405`), `open-mistral-nemo` and , `mistral-large-latest` (`mistral-large-2407`)
12
12
13
13
July 18, 2024
14
-
- We released Mistral NeMo (`open-mistral-nemo`).
14
+
- We released Mistral Nemo (`open-mistral-nemo`).
15
15
16
16
July 16, 2024
17
17
- We released Codestral Mamba (`open-codestral-mamba`) and Mathstral.
| Mistral Large |:heavy_check_mark: <br/> [Mistral Research License](https://mistral.ai/licenses/MRL-0.1.md)|:heavy_check_mark:|Our flagship model with state-of-the-art reasoning, knowledge, and coding capabilities. It's ideal for complex tasks that require large reasoning capabilities or are highly specialized (Synthetic Text Generation, Code Generation, RAG, or Agents). Learn more on our [blog post](https://mistral.ai/news/mistral-large-2407/)| 128k |`mistral-large-latest`|
16
-
| Mistral NeMo|:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:| A 12B model built with the partnership with Nvidia. It is easy to use and a drop-in replacement in any system using Mistral 7B that it supersedes. Learn more on our [blog post](https://mistral.ai/news/mistral-nemo/)| 128k |`open-mistral-nemo`|
13
+
| Model |Weight availability|Available via API| Description | Max Tokens| API Endpoints|Version|
| Mistral Large |:heavy_check_mark: <br/> [Mistral Research License](https://mistral.ai/licenses/MRL-0.1.md)|:heavy_check_mark:|Our flagship model with state-of-the-art reasoning, knowledge, and coding capabilities. It's ideal for complex tasks that require large reasoning capabilities or are highly specialized (Synthetic Text Generation, Code Generation, RAG, or Agents). Learn more on our [blog post](https://mistral.ai/news/mistral-large-2407/)| 128k |`mistral-large-latest`|24.07|
16
+
| Mistral Nemo|:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:| A 12B model built with the partnership with Nvidia. It is easy to use and a drop-in replacement in any system using Mistral 7B that it supersedes. Learn more on our [blog post](https://mistral.ai/news/mistral-nemo/)| 128k |`open-mistral-nemo`|24.07|
17
17
18
18
-**Specialized models**
19
19
20
-
| Model |Available Open-weight|Available via API| Description | Max Tokens| API Endpoints|
| Codestral |:heavy_check_mark: <br/> [Mistral AI Non-Production License](https://mistral.ai/licenses/MNPL-0.1.md)|:heavy_check_mark:| A cutting-edge generative model that has been specifically designed and optimized for code generation tasks, including fill-in-the-middle and code completion. Learn more on our [blog post](https://mistral.ai/news/codestral/)| 32k |`codestral-latest`|
23
-
| Mistral Embeddings||:heavy_check_mark:| A model that converts text into numerical vectors of embeddings in 1024 dimensions. Embedding models enable retrieval and retrieval-augmented generation applications. It achieves a retrieval score of 55.26 on MTEB | 8k |`mistral-embed`|
20
+
| Model |Weight availability|Available via API| Description | Max Tokens| API Endpoints|Version|
| Codestral |:heavy_check_mark: <br/> [Mistral AI Non-Production License](https://mistral.ai/licenses/MNPL-0.1.md)|:heavy_check_mark:| A cutting-edge generative model that has been specifically designed and optimized for code generation tasks, including fill-in-the-middle and code completion. Learn more on our [blog post](https://mistral.ai/news/codestral/)| 32k |`codestral-latest`|24.05|
23
+
| Mistral Embed||:heavy_check_mark:| A model that converts text into numerical vectors of embeddings in 1024 dimensions. Embedding models enable retrieval and retrieval-augmented generation applications. It achieves a retrieval score of 55.26 on MTEB | 8k |`mistral-embed`|23.12|
24
24
25
25
-**Research models**
26
26
27
-
| Model |Available Open-weight|Available via API| Description | Max Tokens| API Endpoints|
| Mistral 7B |:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:|The first dense model released by Mistral AI, perfect for experimentation, customization, and quick iteration. At the time of the release, it matched the capabilities of models up to 30B parameters. Learn more on our [blog post](https://mistral.ai/news/announcing-mistral-7b/)| 32k |`open-mistral-7b`|
30
-
| Mixtral 8x7B |:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:|A sparse mixture of experts model. As such, it leverages up to 45B parameters but only uses about 12B during inference, leading to better inference throughput at the cost of more vRAM. Learn more on the dedicated [blog post](https://mistral.ai/news/mixtral-of-experts/)| 32k |`open-mixtral-8x7b`|
31
-
| Mixtral 8x22B |:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:|A bigger sparse mixture of experts model. As such, it leverages up to 141B parameters but only uses about 39B during inference, leading to better inference throughput at the cost of more vRAM. Learn more on the dedicated [blog post](https://mistral.ai/news/mixtral-8x22b/)| 64k |`open-mixtral-8x22b`|
32
-
| Mathstral |:heavy_check_mark: <br/> Apache2 || A math-specific 7B model designed for math reasoning and scientific tasks. Learn more on our [blog post](https://mistral.ai/news/mathstral/)| 32k | NA|
33
-
| Codestral Mamba |:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:| A Mamba 2 language model specialized in code generation. Learn more on our [blog post](https://mistral.ai/news/codestral-mamba/)| 256k |`open-codestral-mamba`|
27
+
| Model |Weight availability|Available via API| Description | Max Tokens| API Endpoints|Version|
| Mistral 7B |:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:|The first dense model released by Mistral AI, perfect for experimentation, customization, and quick iteration. At the time of the release, it matched the capabilities of models up to 30B parameters. Learn more on our [blog post](https://mistral.ai/news/announcing-mistral-7b/)| 32k |`open-mistral-7b`| v0.3|
30
+
| Mixtral 8x7B |:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:|A sparse mixture of experts model. As such, it leverages up to 45B parameters but only uses about 12B during inference, leading to better inference throughput at the cost of more vRAM. Learn more on the dedicated [blog post](https://mistral.ai/news/mixtral-of-experts/)| 32k |`open-mixtral-8x7b`|v0.1|
31
+
| Mixtral 8x22B |:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:|A bigger sparse mixture of experts model. As such, it leverages up to 141B parameters but only uses about 39B during inference, leading to better inference throughput at the cost of more vRAM. Learn more on the dedicated [blog post](https://mistral.ai/news/mixtral-8x22b/)| 64k |`open-mixtral-8x22b`|v0.1|
32
+
| Mathstral |:heavy_check_mark: <br/> Apache2 || A math-specific 7B model designed for math reasoning and scientific tasks. Learn more on our [blog post](https://mistral.ai/news/mathstral/)| 32k | NA|v0.1|
33
+
| Codestral Mamba |:heavy_check_mark: <br/> Apache2 |:heavy_check_mark:| A Mamba 2 language model specialized in code generation. Learn more on our [blog post](https://mistral.ai/news/codestral-mamba/)| 256k |`open-codestral-mamba`|v0.1|
34
34
35
35
## Pricing
36
36
@@ -67,7 +67,7 @@ It can be used for complex multilingual reasoning tasks, including text understa
67
67
-[Codestral](https://mistral.ai/news/codestral/): as a 22B model, Codestral sets a new standard on the performance/latency space for code generation compared to previous models used for coding.
68
68
-[Codestral-Mamba](https://mistral.ai/news/codestral-mamba/): we have trained this model with advanced code and reasoning capabilities, enabling the model to have a strong performance on par with SOTA transformer-based models.
69
69
-[Mathstral](https://mistral.ai/news/mathstral/): Mathstral stands on the shoulders of Mistral 7B and specialises in STEM subjects. It achieves state-of-the-art reasoning capacities in its size category across various industry-standard benchmarks.
70
-
-[Mistral NeMo](https://mistral.ai/news/mistral-nemo/): Mistral NeMo's reasoning, world knowledge, and coding performance are state-of-the-art in its size category. As it relies on standard architecture, Mistral NeMo is easy to use and a drop-in replacement in any system using Mistral 7B that it supersedes.
70
+
-[Mistral Nemo](https://mistral.ai/news/mistral-nemo/): Mistral Nemo's reasoning, world knowledge, and coding performance are state-of-the-art in its size category. As it relies on standard architecture, Mistral Nemo is easy to use and a drop-in replacement in any system using Mistral 7B that it supersedes.
71
71
72
72
73
73
## Picking a model
@@ -86,7 +86,7 @@ Today, Mistral models are behind many LLM applications at scale. Here is a brief
86
86
When selecting a model, it is essential to evaluate the performance, and cost trade-offs. Depending on what’s most important for your application, your choice may differ significantly. Note that the models will be updated over time, the information we share below only reflect the current state of the models.
87
87
88
88
In general, the larger the model, the better the performance. For instance, when looking at the popular benchmark MMLU (Massive Multitask Language Understanding), the performance ranking of Mistral’s models is as follows:
0 commit comments