Provider: Qwen LM
License: Custom Qwen License (for research and limited commercial use)
Access: Hosted on Kaggle Models and other compatible platforms
Architecture: Decoder-only transformer (Qwen-family design)
Parameters: 32 billion
π Overview
QWQ-32B is a flagship model in the Qwen LM series, designed for high-quality multilingual instruction following. It builds on the Qwen family architecture and targets advanced reasoning, dialog, and code generation across languages.
Key highlights:
- π Multilingual: Supports instruction following and conversation in multiple languages
- π§ High-Performance Core: Tuned for both reasoning and coding tasks
- πͺ Chat Optimized: Works well for assistant-style deployments, especially in long-form or multi-turn exchanges
βοΈ Technical Details
- Architecture: Transformer decoder
- Parameters: 32B
- Tokenizer: Based on Qwen tokenizer for multilingual efficiency
- Training Data: Not fully disclosed; likely includes code, chat, multilingual corpora
- Context Length: Extended (exact size not listed)
π Deployment
- Kaggle Repo: QWQ-32B on Kaggle
- API Compatibility: Hugging Face Transformers, vLLM, and Qwen’s ecosystem tools
- Inference: Supports efficient quantized serving and long-context setups