X

Qwen 3

Information

Qwen 3 is the latest large reasoning model developed by Alibaba company. It surpass multiple baselines on coding, math and surpass SOTA model performance on multiple benchmarks. It is said to be released by May, 2025. # Qwen3

Qwen Chat   |    Hugging Face | ModelScope   | Paper | BlogDocumentation
Demo   | WeChat (微信)   | Discord  

Visit our Hugging Face or ModelScope organization (click links above), search checkpoints with names starting with Qwen3 or visit the [Qwen3 collection](https://huggingface.co/collections/Qwen/qwen3-67dd247413f0e2e4f653967f), and you will find all you need! Enjoy! - Quickstart: the basic usages and demonstrations; - Inference: the guidance for the inference with Transformers, including batch inference, streaming, etc.; - Run Locally: the instructions for running LLM locally on CPU and GPU, with frameworks like llama.cpp and Ollama; - Deployment: the demonstration of how to deploy Qwen for large-scale inference with frameworks like SGLang, vLLM, TGI, etc.; - Quantization: the practice of quantizing LLMs with GPTQ, AWQ, as well as the guidance for how to make high-quality quantized GGUF files; - Training: the instructions for post-training, including SFT and RLHF (TODO) with frameworks like Axolotl, LLaMA-Factory, etc. - Framework: the usage of Qwen with frameworks for application, e.g., RAG, Agent, etc. ## Introduction We are excited to announce the release of Qwen3, the latest addition to the Qwen family of large language models. These models represent our most advanced and intelligent systems to date, improving from our experience in building QwQ and Qwen2.5. We are making the weights of Qwen3 available to the public, including both dense and Mixture-of-Expert (MoE) models. The highlights from Qwen3 include: - **Dense and Mixture-of-Experts (MoE) models of various sizes**, available in 0.6B, 1.7B, 4B, 8B, 14B, 32B and 30B-A3B, 235B-A22B. - **Seamless switching between thinking mode** (for complex logical reasoning, math, and coding) and **non-thinking mode** (for efficient, general-purpose chat), ensuring optimal performance across various scenarios. - **Significantly enhancement in reasoning capabilities**, surpassing previous QwQ (in thinking mode) and Qwen2.5 instruct models (in non-thinking mode) on mathematics, code generation, and commonsense logical reasoning. - **Superior human preference alignment**, excelling in creative writing, role-playing, multi-turn dialogues, and instruction following, to deliver a more natural, engaging, and immersive conversational experience. - **Expertise in agent capabilities**, enabling precise integration with external tools in both thinking and unthinking modes and achieving leading performance among open-source models in complex agent-based tasks. - **Support of 100+ languages and dialects** with strong capabilities for **multilingual instruction following** and **translation**.

Prompts

1

I have a project name for example "project_a" and I want to support both python (pypi) and typescript (npm) services. Additionally, I have some front end plugin which is associated with the APIs (GET). The package support various endpoint and registry service. How can I set the package folder?

2

黄晓明->叶柯 王宝强->

Reviews

Tags


  • AILearner98 2025-05-12 22:54
    Interesting:5,Helpfulness:5,Correctness:5
    Prompt: I have a project name for example "project_a" and I want to support both python (pypi) and typescript (npm) services. Additionally, I have some front end plugin which is associated with the APIs (GET). The package support various endpoint and registry service. How can I set the package folder?

    I asked Qwen3 to help me with the coding problem, which is to create a package folder structure for both python and typescript. It should also contains a folder for plugin. Right now. Qwen3 provides the best answer to me compared to DeepSeek and many other.


  • Ogino Chihiro 2025-04-29 10:01
    Interesting:5,Helpfulness:5,Correctness:5
    Prompt: 黄晓明->叶柯 王宝强->

    The logic question from Qwen3 is correct. The relation is the divorce and current girlfriend relationship. It successfully answered Feng Qing. Besides, the speed genration of first token is really fast.


  • ai4science03 2025-04-28 23:16
    Interesting:5,Helpfulness:5,Correctness:5

    Qwen 2.5 models have already been widely adopted on many industrial applications. And Qwen 3 release will be even more exciting and I think it will set new standard of LLM and even the large reasoning models.

Write Your Review

Detailed Ratings

ALL
Correctness
Helpfulness
Interesting
Upload Pictures and Videos

Name
Size
Type
Last Modified