qwen/qwen3-max-2026-01-23

Qwen3-Max is a flagship large language model designed for ultra-long context understanding, powerful reasoning, and high-performance text and code generation, making it well suited for complex, large-scale, and production-grade AI applications.

LLMFP8NEW
Qwen3 Max 20260123
LLM

Qwen3-Max is a flagship large language model designed for ultra-long context understanding, powerful reasoning, and high-performance text and code generation, making it well suited for complex, large-scale, and production-grade AI applications.

Parametreler

Kod örneği

import os
from openai import OpenAI

client = OpenAI(
    api_key=os.getenv("ATLASCLOUD_API_KEY"),
    base_url="https://api.atlascloud.ai/v1"
)

response = client.chat.completions.create(
    model="qwen/qwen3-max-2026-01-23",
    messages=[
    {
        "role": "user",
        "content": "hello"
    }
],
    max_tokens=1024,
    temperature=0.7
)

print(response.choices[0].message.content)

Kurulum

Programlama diliniz için gerekli paketi kurun.

bash
pip install requests

Kimlik Doğrulama

Tüm API istekleri, API anahtarı ile kimlik doğrulama gerektirir. API anahtarınızı Atlas Cloud kontrol panelinden alabilirsiniz.

bash
export ATLASCLOUD_API_KEY="your-api-key-here"

HTTP Başlıkları

python
import os

API_KEY = os.environ.get("ATLASCLOUD_API_KEY")
headers = {
    "Content-Type": "application/json",
    "Authorization": f"Bearer {API_KEY}"
}
API anahtarınızı güvende tutun

API anahtarınızı asla istemci tarafı kodunda veya herkese açık depolarda ifşa etmeyin. Bunun yerine ortam değişkenleri veya arka uç proxy kullanın.

İstek gönder

import requests

url = "https://api.atlascloud.ai/v1/chat/completions"
headers = {
    "Content-Type": "application/json",
    "Authorization": "Bearer $ATLASCLOUD_API_KEY"
}
data = {
    "model": "your-model",
    "messages": [{"role": "user", "content": "Hello"}],
    "max_tokens": 1024
}

response = requests.post(url, headers=headers, json=data)
print(response.json())

Input Schema

İstek gövdesinde aşağıdaki parametreler kabul edilir.

Toplam: 9Zorunlu: 2İsteğe Bağlı: 7
modelstringrequired
The model ID to use for the completion.
Example: "qwen/qwen3-max-2026-01-23"
messagesarray[object]required
A list of messages comprising the conversation so far.
rolestringrequired
The role of the message author. One of "system", "user", or "assistant".
systemuserassistant
contentstringrequired
The content of the message.
max_tokensinteger
The maximum number of tokens to generate in the completion.
Default: 1024Min: 1
temperaturenumber
Sampling temperature between 0 and 2. Higher values make output more random, lower values more focused and deterministic.
Default: 0.7Min: 0Max: 2
top_pnumber
Nucleus sampling parameter. The model considers the tokens with top_p probability mass.
Default: 1Min: 0Max: 1
streamboolean
If set to true, partial message deltas will be sent as server-sent events.
Default: false
stoparray[string]
Up to 4 sequences where the API will stop generating further tokens.
frequency_penaltynumber
Penalizes new tokens based on their existing frequency in the text so far. Between -2.0 and 2.0.
Default: 0Min: -2Max: 2
presence_penaltynumber
Penalizes new tokens based on whether they appear in the text so far. Between -2.0 and 2.0.
Default: 0Min: -2Max: 2

Örnek İstek Gövdesi

json
{
  "model": "qwen/qwen3-max-2026-01-23",
  "messages": [
    {
      "role": "user",
      "content": "Hello"
    }
  ],
  "max_tokens": 1024,
  "temperature": 0.7,
  "stream": false
}

Output Schema

API, ChatCompletion uyumlu bir yanıt döndürür.

idstringrequired
Unique identifier for the completion.
objectstringrequired
Object type, always "chat.completion".
Default: "chat.completion"
createdintegerrequired
Unix timestamp of when the completion was created.
modelstringrequired
The model used for the completion.
choicesarray[object]required
List of completion choices.
indexintegerrequired
Index of the choice.
messageobjectrequired
The generated message.
finish_reasonstringrequired
The reason generation stopped.
stoplengthcontent_filter
usageobjectrequired
Token usage statistics.
prompt_tokensintegerrequired
Number of tokens in the prompt.
completion_tokensintegerrequired
Number of tokens in the completion.
total_tokensintegerrequired
Total tokens used.

Örnek Yanıt

json
{
  "id": "chatcmpl-abc123",
  "object": "chat.completion",
  "created": 1700000000,
  "model": "model-name",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "Hello! How can I assist you today?"
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 10,
    "completion_tokens": 20,
    "total_tokens": 30
  }
}

Atlas Cloud Skills

Atlas Cloud Skills, 300'den fazla AI modelini doğrudan AI kodlama asistanınıza entegre eder. Kurmak için tek bir komut, ardından görüntü, video oluşturmak ve LLM ile sohbet etmek için doğal dil kullanın.

Desteklenen İstemciler

Claude Code
OpenAI Codex
Gemini CLI
Cursor
Windsurf
VS Code
Trae
GitHub Copilot
Cline
Roo Code
Amp
Goose
Replit
40+ desteklenen i̇stemciler

Kurulum

bash
npx skills add AtlasCloudAI/atlas-cloud-skills

API Anahtarını Ayarla

API anahtarınızı Atlas Cloud kontrol panelinden alın ve ortam değişkeni olarak ayarlayın.

bash
export ATLASCLOUD_API_KEY="your-api-key-here"

Yetenekler

Kurulduktan sonra, tüm Atlas Cloud modellerine erişmek için AI asistanınızda doğal dil kullanabilirsiniz.

Görüntü OluşturmaNano Banana 2, Z-Image ve daha fazla model ile görüntüler oluşturun.
Video OluşturmaKling, Vidu, Veo vb. ile metin veya görüntülerden videolar oluşturun.
LLM SohbetQwen, DeepSeek ve diğer büyük dil modelleri ile sohbet edin.
Medya YüklemeGörüntü düzenleme ve görüntüden videoya iş akışları için yerel dosyaları yükleyin.

MCP Server

Atlas Cloud MCP Server, IDE'nizi Model Context Protocol aracılığıyla 300'den fazla AI modeline bağlar. Herhangi bir MCP uyumlu istemci ile çalışır.

Desteklenen İstemciler

Cursor
VS Code
Windsurf
Claude Code
OpenAI Codex
Gemini CLI
Cline
Roo Code
100+ desteklenen i̇stemciler

Kurulum

bash
npx -y atlascloud-mcp

Yapılandırma

Aşağıdaki yapılandırmayı IDE'nizin MCP ayarları dosyasına ekleyin.

json
{
  "mcpServers": {
    "atlascloud": {
      "command": "npx",
      "args": [
        "-y",
        "atlascloud-mcp"
      ],
      "env": {
        "ATLASCLOUD_API_KEY": "your-api-key-here"
      }
    }
  }
}

Mevcut Araçlar

atlas_generate_imageMetin istemlerinden görüntüler oluşturun.
atlas_generate_videoMetin veya görüntülerden videolar oluşturun.
atlas_chatBüyük dil modelleri ile sohbet edin.
atlas_list_models300'den fazla mevcut AI modelini keşfedin.
atlas_quick_generateOtomatik model seçimi ile tek adımda içerik oluşturma.
atlas_upload_mediaAPI iş akışları için yerel dosyaları yükleyin.

QwQ-32B

Introduction

QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.

Image 5

This repo contains the QwQ 32B model, which has the following features:

  • Type: Causal Language Models
  • Training Stage: Pretraining & Post-training (Supervised Finetuning and Reinforcement Learning)
  • Architecture: transformers with RoPE, SwiGLU, RMSNorm, and Attention QKV bias
  • Number of Parameters: 32.5B
  • Number of Paramaters (Non-Embedding): 31.0B
  • Number of Layers: 64
  • Number of Attention Heads (GQA): 40 for Q and 8 for KV
  • Context Length: Full 131,072 tokens
    • For prompts exceeding 8,192 tokens in length, you must enable YaRN as outlined in this section.

Note: For the best experience, please review the usage guidelines before deploying QwQ models.

You can try our demo or access QwQ models via QwenChat.

For more details, please refer to our blog, GitHub, and Documentation.

Requirements

QwQ is based on Qwen2.5, whose code has been in the latest Hugging face transformers. We advise you to use the latest version of transformers.

With transformers<4.37.0, you will encounter the following error:

KeyError: 'qwen2'

Quickstart

Here provides a code snippet with apply_chat_template to show you how to load the tokenizer and model and how to generate contents.

from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "Qwen/QwQ-32B" model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype="auto", device_map="auto" ) tokenizer = AutoTokenizer.from_pretrained(model_name) prompt = "How many r's are in the word \"strawberry\"" messages = [ {"role": "user", "content": prompt} ] text = tokenizer.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) model_inputs = tokenizer([text], return_tensors="pt").to(model.device) generated_ids = model.generate( **model_inputs, max_new_tokens=32768 ) generated_ids = [ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids) ] response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0] print(response)

Usage Guidelines

To achieve optimal performance, we recommend the following settings:

  1. Enforce Thoughtful Output: Ensure the model starts with "\n" to prevent generating empty thinking content, which can degrade output quality. If you use apply_chat_template and set add_generation_prompt=True, this is already automatically implemented, but it may cause the response to lack the tag at the beginning. This is normal behavior.

  2. Sampling Parameters:

* Use Temperature=0.6, TopP=0.95, MinP=0 instead of Greedy decoding to avoid endless repetitions. * Use TopK between 20 and 40 to filter out rare token occurrences while maintaining the diversity of the generated output. * For supported frameworks, you can adjust the `presence_penalty` parameter between 0 and 2 to reduce endless repetitions. However, using a higher value may result in occasional language mixing and a slight decrease in performance.
  1. No Thinking Content in History: In multi-turn conversations, the historical model output should only include the final output part and does not need to include the thinking content. This feature is already implemented in apply_chat_template.

  2. Standardize Output Format: We recommend using prompts to standardize model outputs when benchmarking.

* **Math Problems**: Include "Please reason step by step, and put your final answer within \boxed{}." in the prompt. * **Multiple-Choice Questions**: Add the following JSON structure to the prompt to standardize responses: "Please show your choice in the `answer` field with only the choice letter, e.g.,`\"answer\": \"C\"`." in the prompt.
  1. Handle Long Inputs: For inputs exceeding 8,192 tokens, enable YaRN to improve the model's ability to capture long-sequence information effectively.

For supported frameworks, you could add the following to config.json to enable YaRN:

{ ..., "rope_scaling": { "factor": 4.0, "original_max_position_embeddings": 32768, "type": "yarn" } }

For deployment, we recommend using vLLM. Please refer to our Documentation for usage if you are not familar with vLLM. Presently, vLLM only supports static YARN, which means the scaling factor remains constant regardless of input length, potentially impacting performance on shorter texts. We advise adding the rope_scaling configuration only when processing long contexts is required.

Evaluation & Performance

Detailed evaluation results are reported in this 📑 blog.

For requirements on GPU memory and the respective throughput, see results here.

Citation

If you find our work helpful, feel free to give us a cite.

@misc{qwq32b, title = {QwQ-32B: Embracing the Power of Reinforcement Learning}, url = {https://qwenlm.github.io/blog/qwq-32b/}, author = {Qwen Team}, month = {March}, year = {2025} } @article{qwen2.5, title={Qwen2.5 Technical Report}, author={An Yang and Baosong Yang and Beichen Zhang and Binyuan Hui and Bo Zheng and Bowen Yu and Chengyuan Li and Dayiheng Liu and Fei Huang and Haoran Wei and Huan Lin and Jian Yang and Jianhong Tu and Jianwei Zhang and Jianxin Yang and Jiaxi Yang and Jingren Zhou and Junyang Lin and Kai Dang and Keming Lu and Keqin Bao and Kexin Yang and Le Yu and Mei Li and Mingfeng Xue and Pei Zhang and Qin Zhu and Rui Men and Runji Lin and Tianhao Li and Tianyi Tang and Tingyu Xia and Xingzhang Ren and Xuancheng Ren and Yang Fan and Yang Su and Yichang Zhang and Yu Wan and Yuqiong Liu and Zeyu Cui and Zhenru Zhang and Zihan Qiu}, journal={arXiv preprint arXiv:2412.15115}, year={2024} }

300+ Model ile Başlayın,

Tüm modelleri keşfet