Loading AI tools
French artificial intelligence company From Wikipedia, the free encyclopedia
Mistral AI, headquartered in Paris, France specializes in artificial intelligence (AI) products and focuses on open-weight large language models,[1][2] (LLMs). Founded in April 2023 by former engineers from Google DeepMind[3] and Meta Platforms, the company has gained prominence as an alternative to proprietary AI systems. Named after the mistral – a powerful, cold wind in southern France[4] – the company emphasized openness and innovation in the AI field. Mistral AI positions itself as an alternative to proprietary models.[5]
Company type | Private |
---|---|
Industry | Artificial intelligence |
Founded | 28 April 2023 |
Founders |
|
Headquarters | Paris , France |
Products |
|
Website | mistral |
In October 2023, Mistral AI raised €385 million.[6] By December 2023, it was valued at over $2 billion.[7][8][9]
In June 2024, Mistral AI secured a €600 million ($645 million) founding round, elevating its valuation to €5.8 billion ($6.2 billion).[10] Led by venture capital firm General Catalyst,[11] this round resulted in additional contributions from existing investors. The funds aim to support the company's expansion.
Mistral AI has published three open-source models available as weights.[12] Additionally, three more models – Small, Medium, and Large – are available via API only.[13][14]
Based on valuation, the company is in fourth place in the global AI race and in first place outside the San Francisco Bay Area, ahead of several of its peers, such as Cohere, Hugging Face, Inflection, Perplexity and Together.[15] Mistral AI aims to "democratize" AI by focusing on open-source innovation.[16]
Mistral AI was established in April 2023 by three French AI researchers: Arthur Mensch, Guillaume Lample and Timothée Lacroix.[17] Mensch, a former researcher at Google DeepMind, brought expertise in advanced AI systems, while Lample and Lacroix contributed their experience from Meta Platforms,[18] where they specialized in developing large-scale AI models. The trio initially met during their studies at École Polytechnique,[4] a public university in France.
In June 2023, the start-up carried out a first fundraising of €105 million ($117 million) with investors including the American fund Lightspeed Venture Partners, Eric Schmidt, Xavier Niel and JCDecaux. The valuation is then estimated by the Financial Times at €240 million ($267 million).
On 27 September 2023, the company made its language processing model “Mistral 7B” available under the free Apache 2.0 license. This model has 7 billion parameters, a small size compared to its competitors.
On 10 December 2023, Mistral AI announced that it had raised €385 million ($428 million) as part of its second fundraising. This round of financing involves the Californian fund Andreessen Horowitz, BNP Paribas and the software publisher Salesforce.[19]
On 11 December 2023, the company released the Mixtral 8x7B model with 46.7 billion parameters but using only 12.9 billion per token with mixture of experts architecture. The model masters 5 languages (French, Spanish, Italian, English and German) and outperforms, according to its developers' tests, the "LLama 2 70B" model from Meta. A version trained to follow instructions and called “Mixtral 8x7B Instruct” is also offered.[20]
On 26 February 2024, Microsoft announced a new partnership with the company to expand its presence in the artificial intelligence industry. Under the agreement, Mistral's language models will be available on Microsoft's Azure cloud, while the multilingual conversational assistant Le Chat will be launched in the style of ChatGPT.[21]
On 10 April 2024, the company released the mixture of expert models, Mixtral 8x22B, offering high performance on various benchmarks compared to other open models.[22]
On 16 April 2024, reporting revealed that Mistral was in talks to raise €500 million, a deal that would more than double its current valuation to at least €5 billion.[23]
On November 19, 2024, the company announced updates for Le Chat. It added the ability to create images, in partnership with Black Forest Labs, utilizing the Flux Pro model. Additionally, it introduced the capability to search for information on the internet to provide reliable and up-to-date information. Furthermore, it launched the Canvas system, a collaborative interface where the AI generates code and the user can modify it. The company also introduced a new model, Pixtral Large, which is an improvement over Pixtral 12B, integrating a 1-billion-parameter visual encoder coupled with Mistral Large 2. This model has also been enhanced, particularly for long contexts and function calls.[24]
The company had over 100 employees by late fall 2024.
Mistral 7B is a 7.3B parameter language model using the transformers architecture. Officially released on September 27, 2023, via a BitTorrent magnet link,[25] and Hugging Face.[26] The model was released under the Apache 2.0 license. The release blog post claimed the model outperforms LLaMA 2 13B on all benchmarks tested, and is on par with LLaMA 34B on many benchmarks tested.[27]
Mistral 7B employs grouped-query attention (GQA), which is a variant of the standard attention mechanism. This architecture optimizes performance by calculating attention within specific groups of hidden states rather than across all hidden states, improving efficiency and scalability.[28]
Both a base model and "instruct" model were released with the latter receiving additional tuning to follow chat-style prompts. The fine-tuned model is only intended for demonstration purposes, and does not have guardrails or moderation built-in.[27]
Much like Mistral's first model, Mixtral 8x7B was released via a BitTorrent link posted on Twitter on December 9, 2023,[1] and later Hugging Face and a blog post were released two days later.[20]
Unlike the previous Mistral model, Mixtral 8x7B uses a sparse mixture of experts architecture. The model has 8 distinct groups of "experts", giving the model a total of 46.7B usable parameters.[29][30] Each single token can only use 12.9B parameters, therefore giving the speed and cost that a 12.9B parameter model would incur.[20]
Mistral AI's testing shows the model beats both LLaMA 70B, and GPT-3.5 in most benchmarks.[31]
In March 2024, research conducted by Patronus AI comparing performance of LLMs on a 100-question test with prompts to generate text from books protected under U.S. copyright law found that Open AI's GPT-4, Mixtral, Meta AI's LLaMA-2, and Anthropic's Claude 2 generated copyrighted text verbatim in 44%, 22%, 10%, and 8% of responses respectively.[32][33]
Similar to Mistral's previous open models, Mixtral 8x22B was released via a BitTorrent link on Twitter on April 10, 2024,[34] with a release on Hugging Face soon after.[35] The model uses an architecture similar to that of Mistral 8x7B, but with each expert having 22 billion parameters instead of 7. In total, the model contains 141 billion parameters, as some parameters are shared among the experts.[35]
Mistral Large 2 was announced on July 24, 2024, and released on Hugging Face. Unlike the previous Mistral Large, this version was released with open weights. It is available for free with a Mistral Research Licence, and with a commercial licence for commercial purposes. Mistral AI claims that it is fluent in dozens of languages, including many programming languages. The model has 123 billion parameters and a context length of 128,000 tokens. Its performance in benchmarks is competitive with Llama 3.1 405B, particularly in programming-related tasks.[36][37]
Codestral is Mistral's first code focused open weight model. Codestral was launched on 29 May 2024. It is a lightweight model specifically built for code generation tasks. As of its release date, this model surpasses Meta's Llama3 70B and DeepSeek Coder 33B (78.2% - 91.6%), another code-focused model on the HumanEval FIM benchmark.[38] Mistral claims Codestral is fluent in more than 80 programming languages[39] Codestral has its own license which forbids the usage of Codestral for commercial purposes.[40]
Mathstral 7B
Mathstral 7B is a model with 7 billion parameters released by Mistral AI on July 16, 2024. It focuses on STEM subjects, achieving a score of 56.6% on the MATH benchmark and 63.47% on the MMLU benchmark.[41] The model was produced in collaboration with Project Numina,[42] and was released under the Apache 2.0 License. It has a context length of 32k tokens.[41]
Codestral Mamba 7B
Codestral Mamba is based on the Mamba 2 architecture, which allows it to generate responses even with longer input.[42] Unlike Codestral, it was released under the Apache 2.0 license. While previous releases often included both the base model and the instruct version, only the instruct version of Codestral Mamba was released.[43]
Unlike Mistral 7B, Mixtral 8x7B and Mixtral 8x22B, the following models are closed-source and only available through the Mistral API.[44]
Mistral Large was launched on February 26, 2024, and Mistral claims it is second in the world only to OpenAI's GPT-4.
It is fluent in English, French, Spanish, German, and Italian, with Mistral claiming understanding of both grammar and cultural context, and provides coding capabilities. As of early 2024, it is Mistral's flagship AI.[45] It is also available on Microsoft Azure.
In July 2024, Mistral Large 2 was released, replacing the original Mistral Large.[46] Unlike the original model, it was released with open weights.[37]
Mistral Medium is trained in various languages including English, French, Italian, German, Spanish and code with a score of 8.6 on MT-Bench.[47] It is ranked in performance above Claude and below GPT-4 on the LMSys ELO Arena benchmark.[48]
The number of parameters, and architecture of Mistral Medium is not known as Mistral has not published public information about it.
Like the Large model, Mistral Small was launched on February 26, 2024.
Seamless Wikipedia browsing. On steroids.
Every time you click a link to Wikipedia, Wiktionary or Wikiquote in your browser's search results, it will show the modern Wikiwand interface.
Wikiwand extension is a five stars, simple, with minimum permission required to keep your browsing private, safe and transparent.