Qwen2.5-Coder-32b-Instruct
★ featuredqwen/qwen2.5-coder-32b-instruct
Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). As of now, Qwen2.5-Coder has covered six mainstream model sizes, 0.5, 1.5, 3, 7, 14, 32 billion parameters, to meet the needs of different developers. Qwen2.5-Coder brings the following improvements upon CodeQwen1.5:
$0.660 in · $1.00 out / 1M32,768 ctx
alibaba/qwen3.5-397b-a17b
Alibaba's Qwen 3.5 is a 397B-parameter mixture-of-experts model with 17B active parameters, offering strong reasoning capabilities with efficient inference.
$0.600 in · $3.60 out / 1M262,144 ctx
qwen/qwen3-30b-a3b-fp8
Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models. Built upon extensive training, Qwen3 delivers groundbreaking advancements in reasoning, instruction-following, agent capabilities, and multilingual support.
$0.051 in · $0.340 out / 1M32,768 ctx
qwen/qwen1.5-14b-chat-awq
Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.
$0.120 in · $0.240 out / 1M4,096 ctx
qwen/qwen1.5-7b-chat-awq
Qwen1.5 is the improved version of Qwen, the large language model series developed by Alibaba Cloud. AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization.
$0.060 in · $0.120 out / 1M4,096 ctx