IBM introduces Open-Source Mistral AI Model, Now Accessible on WatsonX


March 6, 2024 by our News Team

IBM introduces Mixtral-8x7B large language model on Watsonx AI platform for enhanced throughput and performance.

  • Increased throughput by 50% compared to regular model
  • Potential to reduce latency by 35-75%
  • Unique combination of Sparse modeling and Mixture-of-Experts technique


IBM has made an exciting announcement in the world of artificial intelligence (AI) with the availability of the Mixtral-8x7B large language model (LLM) on its Watsonx AI and data platform. Developed by Mistral AI, this popular open-source model is set to expand IBM’s capabilities and provide clients with innovative solutions.

What makes the optimized version of Mixtral-8x7B particularly impressive is its ability to increase throughput by 50 percent compared to the regular model. Through internal testing, IBM has found that this boost in performance could potentially reduce Latency by 35-75 percent, depending on batch size. This is achieved through a process called quantization, which effectively reduces the model size and memory requirements, resulting in faster processing times and lower costs.

The addition of Mixtral-8x7B aligns with IBM’s open, multi-model strategy, which aims to meet clients where they are and offer them choice and flexibility when it comes to scaling enterprise AI solutions. By collaborating with Meta and Hugging Face, as well as partnering with other model leaders, IBM continues to expand its watsonx.ai model catalog. This means clients can now access new capabilities, languages, and modalities to empower their businesses.

IBM’s enterprise-ready foundation models, combined with the watsonx AI and data platform, enable clients to harness the power of generative AI and unlock new insights and efficiencies. Additionally, this allows them to create innovative business models based on principles of trust. With the ability to select the right model for specific use cases and price-performance goals, such as in the finance industry, clients are given the tools they need to drive innovation in their respective domains.

Mixtral-8x7B stands out due to its unique combination of Sparse modeling and the Mixture-of-Experts technique. Sparse modeling focuses on identifying and utilizing only the most essential parts of data, resulting in more efficient models. The Mixture-of-Experts technique, on the other hand, combines different models that specialize in solving different aspects of a problem. This combination allows Mixtral-8x7B to rapidly process and analyze vast amounts of data, providing valuable context-relevant insights.

“Clients are asking for choice and flexibility to deploy models that best suit their unique use cases and business requirements,” said Kareem Yusuf, Senior Vice President of Product Management & Growth at IBM Software. By offering Mixtral-8x7B and other models on the watsonx platform, IBM is not only meeting this demand but also fostering an ecosystem of AI builders and business leaders. This empowers them with the necessary tools and technologies to drive innovation across diverse industries and domains.

In addition to Mixtral-8x7B, IBM has also announced the availability of ELYZA-japanese-Llama-2-7b, a Japanese LLM model open-sourced by ELYZA Corporation, on watsonx. Furthermore, IBM offers Meta’s open-source models Llama-2-13B-chat and Llama-2-70B-chat, along with other third-party models. And there’s more to come in the next few months, as IBM continues to expand its products on the watsonx platform.

IBM introduces Open-Source Mistral AI Model, Now Accessible on WatsonX

About Our Team

Our team comprises industry insiders with extensive experience in computers, semiconductors, games, and consumer electronics. With decades of collective experience, we’re committed to delivering timely, accurate, and engaging news content to our readers.

Background Information


About IBM:

IBM, or International Business Machines Corporation, is a globally American multinational technology company with a storied history dating back to its founding in 1911. Over the decades, IBM has consistently been at the forefront of innovation in the field of information technology. The company is known for its pioneering work in computer hardware, software, and services, with breakthroughs like the IBM System/360 and the invention of the relational database.

IBM website
Latest Articles about IBM

Technology Explained


Latency: Technology latency is the time it takes for a computer system to respond to a request. It is an important factor in the performance of computer systems, as it affects the speed and efficiency of data processing. In the computer industry, latency is a major factor in the performance of computer networks, storage systems, and other computer systems. Low latency is essential for applications that require fast response times, such as online gaming, streaming media, and real-time data processing. High latency can cause delays in data processing, resulting in slow response times and poor performance. To reduce latency, computer systems use various techniques such as caching, load balancing, and parallel processing. By reducing latency, computer systems can provide faster response times and improved performance.

Latest Articles about Latency

LLM: A Large Language Model (LLM) is a highly advanced artificial intelligence system, often based on complex architectures like GPT-3.5, designed to comprehend and produce human-like text on a massive scale. LLMs possess exceptional capabilities in various natural language understanding and generation tasks, including answering questions, generating creative content, and delivering context-aware responses to textual inputs. These models undergo extensive training on vast datasets to grasp the nuances of language, making them invaluable tools for applications like chatbots, content generation, and language translation.

Latest Articles about LLM




Leave a Reply