



const { OpenAI } = require('openai');
const api = new OpenAI({
baseURL: 'https://api.ai.cc/v1',
apiKey: '',
});
const main = async () => {
const result = await api.chat.completions.create({
model: 'microsoft/WizardLM-2-8x22B',
messages: [
{
role: 'system',
content: 'You are an AI assistant who knows everything.',
},
{
role: 'user',
content: 'Tell me, why is the sky blue?'
}
],
});
const message = result.choices[0].message.content;
console.log(`Assistant: ${message}`);
};
main();
import os
from openai import OpenAI
client = OpenAI(
base_url="https://api.ai.cc/v1",
api_key="",
)
response = client.chat.completions.create(
model="microsoft/WizardLM-2-8x22B",
messages=[
{
"role": "system",
"content": "You are an AI assistant who knows everything.",
},
{
"role": "user",
"content": "Tell me, why is the sky blue?"
},
],
)
message = response.choices[0].message.content
print(f"Assistant: {message}")
-
AI Playground

Test all API models in the sandbox environment before you integrate.
We provide more than 300 models to integrate into your app.


Product Detail
✨ Introducing WizardLM 2-8 (22B): A Cutting-Edge LLM from Microsoft
Developed by Microsoft, WizardLM 2-8 (22B) is a state-of-the-art large language model (LLM) released in April 2024. This version 2.0 model is engineered for exceptional performance in complex tasks such as multilingual conversations, advanced reasoning, and dynamic agent-based interactions. Leveraging an advanced architecture and extensive training, WizardLM 2-8 (22B) delivers high-quality, contextually relevant responses across a multitude of applications.
Basic Information:
- Model Name: WizardLM 2-8 (22B)
- Developer: Microsoft
- Release Date: April 2024
- Version: 2.0
- Model Type: Large Language Model (LLM)
🚀 Key Features & Capabilities of WizardLM 2-8 (22B)
- 141 Billion Parameters: Offers robust performance for understanding and generating sophisticated, human-like text.
- Mixture of Experts (MoE) Architecture: Utilizes a combination of specialized sub-models to optimize efficiency and performance.
- Multilingual Support: Seamlessly engages in conversations across multiple languages, making it ideal for global applications.
- High Efficiency: Achieves competitive performance with remarkable speed, often outperforming larger models.
- Synthetic Training Data: Utilizes a fully AI-powered synthetic training system to significantly enhance its learning capabilities and adaptability.
🎯 Intended Use & Applications
WizardLM 2-8 (22B) is primarily developed for software developers and researchers. It provides advanced Natural Language Processing (NLP) capabilities for integration into various applications, including but not limited to:
- Chatbots and Virtual Assistants: Enhancing conversational AI with more natural and intelligent interactions.
- Content Generation Tools: Automating the creation of diverse textual content.
- Multilingual Applications: Facilitating communication and content processing across different languages.
- Complex Reasoning Systems: Powering applications that require sophisticated logical inference.
⚙️ Technical Deep Dive into WizardLM 2-8 (22B)
Architecture: Mixture of Experts (MoE)
The core of WizardLM 2-8 (22B)'s efficiency and power lies in its Mixture of Experts (MoE) architecture. This design allows the model to dynamically activate specific subsets of its parameters based on the input it receives. This adaptive approach significantly boosts both processing speed and overall performance, making it exceptionally effective for complex reasoning tasks and maintaining high contextual relevance.
Training Data: AI-Powered Synthetic Generation
WizardLM 2-8 (22B) was trained on a vast and diverse dataset primarily composed of synthetic data generated by advanced AI systems. This innovative training methodology enables rapid learning from a broad spectrum of scenarios.
- Data Source & Size: The dataset encompasses a wide array of topics and linguistic styles, ensuring robust and versatile responses.
- Diversity & Bias: While curated to minimize biases and maximize input diversity, the reliance on synthetic data prompts ongoing discussion regarding its applicability to real-world complexities.
Performance Metrics & Comparisons
WizardLM 2-8 (22B) has consistently showcased impressive performance benchmarks. For detailed visualizations of its capabilities and comparisons against other leading models, refer to the following figures:

🛠️ How to Use WizardLM 2-8 (22B)
Code Samples & API Access
The WizardLM 2-8 (22B) model is readily available on the AI/ML API platform. Developers can integrate this powerful LLM into their applications using the provided API.
Access the model via the AI/ML API platform.
<snippet data-name="open-ai.chat-completion" data-model="microsoft/WizardLM-2-8x22B"></snippet>
Comprehensive API Documentation
For detailed instructions, integration guides, and advanced usage scenarios, refer to the official API documentation:
Explore the API Documentation for full details.
🛡️ Ethical AI Development & Guidelines
Microsoft places a strong emphasis on ethical considerations in the development and deployment of its AI models. Transparency regarding the model's capabilities and limitations is paramount, alongside encouraging responsible usage to prevent misuse or the generation of harmful content.
Commitment to Safety: Notably, the initial release of WizardLM underwent a retraction for extensive toxicity testing, underscoring Microsoft's unwavering commitment to ethical AI standards and user safety.
📝 Licensing Information
WizardLM models, including WizardLM 2-8 (22B), are distributed under an open-source license. This license grants extensive rights for both research and commercial usage, provided that users adhere to established ethical standards and guidelines.
Get Started: Secure your WizardLM 2-8 (22B) API access here.
❓ Frequently Asked Questions (FAQ)
Q1: What is WizardLM 2-8 (22B)?
A1: It's a cutting-edge large language model (LLM) developed by Microsoft, released in April 2024, designed for complex tasks like multilingual conversations, advanced reasoning, and agent-based interactions.
Q2: What are its main features?
A2: Key features include 141 Billion Parameters, a Mixture of Experts (MoE) architecture, multilingual support, high efficiency, and training with AI-powered synthetic data.
Q3: How can I access WizardLM 2-8 (22B)?
A3: The model is available on the AI/ML API platform. Developers can find code samples and detailed API documentation on their website.
Q4: Is WizardLM 2-8 (22B) open-source?
A4: Yes, WizardLM models are released under an open-source license that permits both research and commercial usage, adhering to ethical standards.
Q5: What is Microsoft's stance on AI ethics for this model?
A5: Microsoft prioritizes transparency, responsible usage, and safety. The initial release was even retracted for extensive toxicity testing, highlighting their commitment to ethical AI development.
Learn how you can transformyour company with AICC APIs



Log in