



const { OpenAI } = require('openai');
const api = new OpenAI({
baseURL: 'https://api.ai.cc/v1',
apiKey: '',
});
const main = async () => {
const result = await api.chat.completions.create({
model: 'DiscoResearch/DiscoLM-mixtral-8x7b-v2',
messages: [
{
role: 'system',
content: 'You are an AI assistant who knows everything.',
},
{
role: 'user',
content: 'Tell me, why is the sky blue?'
}
],
});
const message = result.choices[0].message.content;
console.log(`Assistant: ${message}`);
};
main();
import os
from openai import OpenAI
client = OpenAI(
base_url="https://api.ai.cc/v1",
api_key="",
)
response = client.chat.completions.create(
model="DiscoResearch/DiscoLM-mixtral-8x7b-v2",
messages=[
{
"role": "system",
"content": "You are an AI assistant who knows everything.",
},
{
"role": "user",
"content": "Tell me, why is the sky blue?"
},
],
)
message = response.choices[0].message.content
print(f"Assistant: {message}")
-
AI Playground

Test all API models in the sandbox environment before you integrate.
We provide more than 300 models to integrate into your app.


Product Detail
💡 Understanding DiscoLM Mixtral 8x7b
DiscoLM Mixtral 8x7b, developed by DiscoResearch and led by Björn Plüster, is a cutting-edge language model released on December 11, 2023. As a V2 iteration, this text generation model employs a highly efficient sparse mixture of experts (MoE) architecture, making it ideal for a broad spectrum of advanced Natural Language Processing (NLP) tasks.
✨ Key Features at a Glance
- ✓ Sparse Mixture of Experts (MoE) Architecture: Harnesses 8 groups of experts, featuring a substantial 46.7 billion total parameters, yet optimizing for efficiency with only 12.9 billion parameters utilized per token during inference.
- ✓ High Performance: Consistently achieves top-tier benchmarks across diverse NLP tasks, ensuring superior results.
- ✓ Multi-Language Proficiency: Seamlessly operates in English, French, Spanish, Italian, and German.
- ✓ Extended Context Length: Supports an impressive context window of up to 32,768 tokens, allowing for deeper comprehension and generation.
💬 Intended Applications
DiscoLM Mixtral 8x7b is engineered for versatility and excels in applications such as:
- ● Text generation and completion
- ● Conversational AI development
- ● Efficient content creation
- ● High-quality language translation
- ● Advanced NLP research initiatives
🔧 Technical Specifications
Architecture Breakdown
The model's core strength lies in its sparse mixture of experts (MoE) architecture. This intelligent design allows the model to activate only a specific subset of its vast parameters for each token, striking an optimal balance between computational efficiency and peak performance. It's built upon the robust Mixtral framework, finely tuned for causal language modeling.
Training Data & Diversity
DiscoLM Mixtral 8x7b underwent fine-tuning on a rich and diverse collection of datasets, including:
- ● Synthia: A synthetic dataset crafted for general NLP tasks.
- ● MethaMathQA: Specifically tailored for mathematical problem-solving.
- ● Capybara: A comprehensive resource for conversational AI development.
While the exact size remains undisclosed, the training data draws from a wide array of sources, bolstering the model's generalization capabilities. The model's knowledge is current as of December 2023. Extensive efforts were made to incorporate diverse datasets to mitigate biases; however, inherent biases common to large language models may still be present.
📈 Performance & Benchmarks
Key Performance Metrics
- ● ARC (25-shot): 67.32
- ● HellaSwag (10-shot): 86.25
- ● MMLU (5-shot): 70.72
- ● TruthfulQA (0-shot): 54.17
- ● Winogrande (5-shot): 80.72
- ● GSM8k (5-shot): 25.09
Competitive Advantage
DiscoLM Mixtral 8x7b consistently outperforms many contemporary models, including Meta's LLama 2 70B, across various benchmarks, highlighting its superior capabilities. Its MoE architecture also ensures efficient inference speed and strong robustness across diverse inputs, topics, and languages.
📜 Usage & Licensing
Code Samples
Ethical Guidelines
Users are encouraged to deploy DiscoLM Mixtral 8x7b responsibly, being mindful of potential biases and ethical implications. This model is primarily intended for research purposes and should strictly not be used for any harmful activities.
Licensing Information
DiscoLM Mixtral 8x7b is released under the Apache 2.0 license, which permits both commercial and non-commercial utilization.
ⓘ Frequently Asked Questions (FAQ)
- Q: What is DiscoLM Mixtral 8x7b?
- A: DiscoLM Mixtral 8x7b is a state-of-the-art text generation language model by DiscoResearch, utilizing a sparse mixture of experts (MoE) architecture for high performance and efficiency in various NLP tasks.
- Q: What languages does DiscoLM Mixtral 8x7b support?
- A: It is proficient in multiple languages, including English, French, Spanish, Italian, and German.
- Q: How does its MoE architecture benefit users?
- A: The Sparse Mixture of Experts architecture allows the model to activate only a subset of its total parameters per token, leading to optimized computational efficiency without compromising on high performance.
- Q: What is the context length of this model?
- A: DiscoLM Mixtral 8x7b supports an extended context length of up to 32,768 tokens.
- Q: Under what license is DiscoLM Mixtral 8x7b released?
- A: The model is released under the Apache 2.0 license, permitting both commercial and non-commercial use.
Learn how you can transformyour company with AICC APIs



Log in