



const fs = require('fs');
const path = require('path');
const axios = require('axios').default;
const api = new axios.create({
baseURL: 'https://api.ai.cc/v1',
headers: { Authorization: 'Bearer ' },
});
const main = async () => {
try {
const response = await api.post('/tts', {
model: 'microsoft/vibevoice-7b',
script: 'Speaker 0: Hello there! Speaker 1: Hi, how are you?',
speakers: [
{ preset: 'Frank [EN]' }
]
});
const responseData = response.data;
const audioUrl = responseData.audio.url;
const fileName = responseData.audio.file_name;
const audioResponse = await api.get(audioUrl, { responseType: 'stream' });
const dist = path.resolve(__dirname, fileName);
const writeStream = fs.createWriteStream(dist);
audioResponse.data.pipe(writeStream);
writeStream.on('close', () => {
console.log('Audio saved to:', dist);
console.log(`Duration: ${responseData.duration} seconds`);
console.log(`Sample rate: ${responseData.sample_rate} Hz`);
});
} catch (error) {
console.error('Error:', error.message);
}
};
main();
import os
import requests
def main():
url = "https://api.ai.cc/v1/tts"
headers = {
"Authorization": "Bearer ",
}
payload = {
"model": "microsoft/vibevoice-7b",
"script": "Speaker 0: Hello there! Speaker 1: Hi, how are you?",
"speakers": [
{ "preset": "Frank [EN]" }
]
}
try:
response = requests.post(url, headers=headers, json=payload)
response.raise_for_status() # Raise an exception for bad status codes
response_data = response.json()
audio_url = response_data["audio"]["url"]
file_name = response_data["audio"]["file_name"]
audio_response = requests.get(audio_url, stream=True)
audio_response.raise_for_status()
dist = os.path.join(os.path.dirname(__file__), file_name)
with open(dist, "wb") as write_stream:
for chunk in audio_response.iter_content(chunk_size=8192):
if chunk:
write_stream.write(chunk)
print("Audio saved to:", dist)
print(f"Duration: {response_data['duration']} seconds")
print(f"Sample rate: {response_data['sample_rate']} Hz")
except requests.exceptions.RequestException as e:
print(f"Error making request: {e}")
except Exception as e:
print(f"Error: {e}")
main()
-
AI Playground

Test all API models in the sandbox environment before you integrate.
We provide more than 300 models to integrate into your app.


Product Detail
✨ VibeVoice 7B is a groundbreaking AI-powered voice synthesis model designed to produce incredibly natural, expressive, and context-aware speech. It's an ideal solution for developers, content creators, and businesses seeking versatile voice capabilities across various sectors, including media, virtual assistants, gaming, education, and accessibility technologies. Leveraging advanced deep neural architectures, VibeVoice 7B offers customizable voice personas enriched with robust emotional nuance and linguistic precision.
Technical Capabilities & Input Flexibility
✅ Model Input Types
VibeVoice 7B supports a variety of input formats including plain text, SSML (Speech Synthesis Markup Language) for detailed speech control, and prosody parameters to fine-tune intonation, pace, and rhythm. This allows for intricate command over voice outputs, perfectly adaptable to diverse scenarios and user preferences.
💭 Input Length & Context Awareness
The model is capable of processing extended conversational inputs while maintaining strong contextual coherence. This makes it exceptionally well-suited for dynamic dialogues, narrative storytelling, and complex multi-turn interactions.
Performance & Output Quality Metrics
-
⏱ Real-Time Speech Generation: Optimized for rapid response, VibeVoice 7B generates high-fidelity speech at near real-time speeds, perfect for interactive applications like live chatbots and virtual personas.
-
🎧 Audio Fidelity: It delivers crystal-clear, studio-level voice outputs with rich tonal texture, natural prosody, and precise phonetic detail. The model's neural vocoder ensures smooth, artifact-free audio synthesis.
-
🎭 Voice Style Variety: Supports a wide range of voice styles, accents, and emotional tones—from cheerful and energetic to calm and professional—empowering brands to forge unique auditory identities.
Model Architecture & Innovations
-
🧩 Hybrid Transformer-Based Design: VibeVoice 7B utilizes a transformer backbone enhanced with attention mechanisms specifically tailored for speech features. This hybrid design excels at capturing long-range linguistic dependencies and prosodic patterns.
-
😍 Emotional & Expressive Modulation: Advanced embedding vectors simulate emotional states and speaker intent, enabling expressive speech synthesis that far surpasses conventional robotic voices.
-
🌍 Robust Training Dataset: Trained on an extensive multilingual dataset covering diverse demographics, accents, and speaking styles, ensuring high adaptability across languages and domains.
Core Features & Usage Scenarios
-
🧑🗨️ Custom Voice Persona Creation: Users can generate personalized voice variants by fine-tuning speech style, pitch, and emotional parameters. This is ideal for interactive voice applications and unique audio content.
-
🌆 Multi-Domain Applications: Widely applicable for audiobook narration, voice-overs in videos and commercials, in-game character voices, accessibility tools for the visually impaired, and advanced conversational AI systems.
💸 API Pricing
-
$0.042 per generated minute – cost-effective and transparent pricing.
Key Use Cases for VibeVoice 7B
-
🤖 Interactive Virtual Assistants & Chatbots: Empower AI characters with rich, believable vocal personalities that adapt tone based on conversation flow, enhancing user engagement.
-
🎥 Media & Entertainment Voice Production: Generate diverse character voices and scenarios without the need for costly studio recording sessions, streamlining production workflows.
-
💻 Accessibility & Assistive Technology: Create natural-sounding screen readers and communication aids that support emotional expression, significantly improving user experience for the visually impaired.
-
📚 Educational Tools: Facilitate language learning and speech therapy applications with clear, expressive pronunciation and customizable pacing, making learning more effective and engaging.
Code Sample
(Note: This is a placeholder for an actual code snippet or API integration example.)
Comparative Analysis with Leading Voice Synthesis Models
🔊 Vs ElevenLabs (ElevenVoice): While ElevenLabs excels in multi-modal input integration and extensive style transfer, VibeVoice 7B differentiates itself with superior emotional expressiveness and suitability for real-time interaction, offering finer granularity in prosody and contextual speech adaptation.
🔊 Vs Google Text-to-Speech: Google’s TTS solutions offer broad language support and robust integration but often prioritize generality. VibeVoice 7B, conversely, provides richer emotional modulation and advanced personalized voice creation capabilities, making it a preferred choice for creative content and brand-specific voice applications.
🔊 Vs Amazon Polly: Amazon Polly is a robust platform for scalable deployments and multilingual support. However, VibeVoice 7B outperforms it in delivering dynamic, expressive tone variations and achieving higher fidelity naturalness, more effectively mimicking human speech nuances.
🔊 Vs Microsoft Azure Speech Service: Azure Speech focuses heavily on enterprise-grade deployment and transcription synergy. VibeVoice 7B’s core strength lies in its ability to dynamically adapt speech expressivity and style, making it exceptionally suited for narrative and conversational user experiences.
Frequently Asked Questions (FAQ)
❓ What makes VibeVoice 7B's speech synthesis studio-quality?
VibeVoice 7B utilizes a sophisticated cascaded diffusion architecture and multi-scale vocoder processing. This ensures exceptional fidelity, naturalness, and comprehensive acoustic characteristics, capturing both broad prosodic patterns and fine-grained vocal nuances.
❓ How does the 7B parameter scale enhance emotional expressiveness?
The expanded 7B parameter budget allows for sophisticated emotional modeling, nuanced prosodic variations, and detailed spectral modeling. It incorporates specialized emotion encoders and advanced pitch/timing control, enabling speech with remarkable emotional depth and vocal quality.
❓ What voice customization features does VibeVoice 7B offer?
Users have precise control over emotional delivery, high-fidelity voice cloning from limited samples, and granular adjustments for pitch, timbre, and speaking characteristics. Advanced features include emotional arc specification for narratives and accent/dialect adaptation.
❓ Can VibeVoice 7B handle complex narrative and dramatic reading tasks?
Yes, the model demonstrates advanced narrative understanding with appropriate pacing, character voice differentiation in dialogues, emotional progression throughout stories, and dramatic interpretation. Its context-aware prosody modeling adapts delivery based on narrative structure.
❓ What professional applications benefit most from VibeVoice 7B?
Professional applications like audiobook production, video game dialogue, animated content, advertising voiceovers, educational content, and virtual assistant interactions significantly benefit from its studio-grade output quality and extensive creative control.
Learn how you can transformyour company with AICC APIs



Log in