qwen-bg
max-ico04
2K
In
Out
max-ico02
Chat
max-ico03
disable
Guanaco (65B)
Access Guanaco-65B API. Guanaco 65B is an open-source chatbot model that rivals ChatGPT 3.5 Turbo, developed using efficient 4-bit QLoRA finetuning.
Free $1 Tokens for New Members
Text to Speech
                                        const { OpenAI } = require('openai');

const api = new OpenAI({
  baseURL: 'https://api.ai.cc/v1',
  apiKey: '',
});

const main = async () => {
  const result = await api.chat.completions.create({
    model: 'togethercomputer/guanaco-65b',
    messages: [
      {
        role: 'system',
        content: 'You are an AI assistant who knows everything.',
      },
      {
        role: 'user',
        content: 'Tell me, why is the sky blue?'
      }
    ],
  });

  const message = result.choices[0].message.content;
  console.log(`Assistant: ${message}`);
};

main();
                                
                                        import os
from openai import OpenAI

client = OpenAI(
    base_url="https://api.ai.cc/v1",
    api_key="",    
)

response = client.chat.completions.create(
    model="togethercomputer/guanaco-65b",
    messages=[
        {
            "role": "system",
            "content": "You are an AI assistant who knows everything.",
        },
        {
            "role": "user",
            "content": "Tell me, why is the sky blue?"
        },
    ],
)

message = response.choices[0].message.content

print(f"Assistant: {message}")
Docs

One API 300+ AI Models

Save 20% on Costs & $1 Free Tokens
  • ico01-1
    AI Playground

    Test all API models in the sandbox environment before you integrate.

    We provide more than 300 models to integrate into your app.

    copy-img02img01
qwenmax-bg
img
Guanaco (65B)

Product Detail

💡 Guanaco-65B: A Leading Open-Source LLM Chatbot

The Guanaco-65B is a sophisticated, 65 billion parameter open-source chatbot model. Released in 2023 by Tim Dettmers, this advanced text-based Large Language Model (LLM) is a testament to efficient finetuning techniques. It was developed by applying 4-bit QLoRA finetuning to the LLaMA base model, utilizing the comprehensive OASST1 dataset.

Guanaco-65B stands out for its exceptional performance, achieving capabilities comparable to top-tier commercial chatbots such as ChatGPT and BARD. This makes high-performance conversational AI more accessible and cost-effective for a broad range of applications.

✅ Key Features & Advantages

  • Competitive Performance: Demonstrates performance comparable to ChatGPT and BARD on prominent benchmarks like Vicuna and OpenAssistant.
  • Open-Source Availability: Freely accessible for local experimentation and deployment, democratizing access to powerful AI.
  • Efficient 4-bit QLoRA Training: Built upon a replicable and highly efficient 4-bit QLoRA finetuning process.
  • Lightweight Adapter Weights: Utilizes compact adapter weights that seamlessly integrate with LLaMA base models.

🎯 Intended Use Cases

The Guanaco-65B is engineered to empower developers and researchers in deploying and experimenting with cutting-edge conversational AI systems. Its versatility makes it an ideal choice for various applications, including:

  • Developing robust open-domain chatbots
  • Crafting task-oriented dialogue systems
  • Enhancing question-answering functionalities
  • Automating text summarization tasks
  • Generating diverse and creative text content

🌐 Multilingual Capabilities

While Guanaco-65B is inherently a multilingual model, its training on the OASST1 dataset, which is heavily biased towards high-resource languages, suggests optimal performance. Therefore, the model is expected to perform best in English and other widely-resourced languages.

⚙️ Technical Specifications

Architecture:

The Guanaco-65B leverages a LoRA (Low-Rank Adaptation) architecture. This involves adding specific adapter weights to all layers of the underlying LLaMA base model. This design allows for highly efficient finetuning, enabling extensive customization while diligently preserving the core capabilities of the base model.

Training Data:

The model was trained using the OASST1 dataset. This dataset is known for its multilingual nature but is skewed towards high-resource languages. Specific details regarding the dataset's exact size and comprehensive diversity are not publicly disclosed.

Knowledge Cutoff:

The precise knowledge cutoff date for Guanaco-65B is not publicly specified. It is generally understood that its knowledge base reflects the information available up to the finalization date of the OASST1 dataset used for its finetuning.

Performance Metrics:

According to documented reports, Guanaco-65B demonstrates remarkable performance, achieving 99.3 percent of ChatGPT-3.5 Turbo's performance on the demanding Vicuna benchmarks. This impressive evaluation was corroborated by both human assessment and analysis by GPT-4.

🚀 API Usage Example

Integrating Guanaco-65B into your applications is designed for ease of use. Here's a common API usage example, illustrating how to interact with the model for chat completions:

# Example API call for chat completion
import openai
client = openai.OpenAI(
    api_key="YOUR_API_KEY",
    base_url="https://api.together.xyz/v1",
)

chat_completion = client.chat.completions.create(
    model="togethercomputer/guanaco-65b",
    messages=[
        {
            "role": "system",
            "content": "You are a helpful assistant.",
        },
        {
            "role": "user",
            "content": "What is the capital of France?",
        },
    ],
    temperature=0.7,
    max_tokens=512,
)

print(chat_completion.choices[0].message.content)

Note: This illustrative snippet assumes compatibility with OpenAI API standards, often supported by platforms like Together AI. For further details on API integration, you might refer to the Together AI blog post on Guanaco-65B.

⚖️ Ethical Use & Licensing Information

Ethical Guidelines:

As an open-source model, no specific ethical guidelines are formally issued for Guanaco-65B by its developer. Therefore, the responsibility for its responsible deployment, consideration of potential misuse, and adherence to ethical AI practices lies solely with the developers and end-users.

License Type:

The Guanaco adapter weights are licensed under the permissive Apache 2.0 License. However, it is critical to understand that the complete utilization of the Guanaco-65B model necessitates access to the underlying LLaMA base model weights, which are governed by more restrictive licensing terms. Users must ensure full compliance with both sets of licensing agreements.

✨ Conclusion

In essence, the Guanaco-65B represents a powerful and accessible open-source chatbot model that effectively competes with established commercial AI offerings like ChatGPT. It not only highlights the remarkable potential and efficiency of 4-bit QLoRA finetuning but also provides an affordable and reproducible pathway for developing and deploying high-performance conversational AI solutions. Its introduction significantly contributes to the broader accessibility of advanced LLM technology.

❓ Frequently Asked Questions (FAQ)

What is Guanaco-65B?

Guanaco-65B is a 65 billion parameter, open-source chatbot model developed by Tim Dettmers. It's built upon the LLaMA base model and finetuned using the 4-bit QLoRA technique, offering performance comparable to leading commercial AI chatbots.

How does Guanaco-65B's performance compare to ChatGPT?

According to documentation and benchmarks, Guanaco-65B achieves 99.3% of ChatGPT-3.5 Turbo's performance on the Vicuna benchmarks, as evaluated by both human raters and GPT-4, demonstrating its highly competitive nature.

What is QLoRA finetuning?

QLoRA (Quantized Low-Rank Adaptation) is an efficient 4-bit quantization method for finetuning large language models. It significantly reduces memory consumption while maintaining high performance, making it feasible to train and deploy massive models on more accessible hardware.

Can Guanaco-65B be used for commercial purposes?

The Guanaco adapter weights are under the Apache 2.0 License, which generally permits commercial use. However, the foundational LLaMA base model weights have more restrictive licensing terms. Users must ensure compliance with both licenses for any commercial application.

What languages are best supported by Guanaco-65B?

While it is a multilingual model, the OASST1 dataset it was trained on is heavily biased towards high-resource languages. Consequently, Guanaco-65B is expected to perform optimally with English and other similar widely-resourced languages.

Learn how you can transformyour company with AICC APIs

Discover how to revolutionize your business with AICC API! Unlock powerfultools to automate processes, enhance decision-making, and personalize customer experiences.
Contact sales
api-right-1
model-bg02-1

One API
300+ AI Models

Save 20% on Costs