Grok-2
The next-generation model from xAI, built on a new architecture and fully integrated into X (formerly Twitter) as part of Elon Musk’s AI assistant efforts.
The next-generation model from xAI, built on a new architecture and fully integrated into X (formerly Twitter) as part of Elon Musk’s AI assistant efforts.
Meta’s 7B-parameter base language model from the LLaMA 2 series, designed for general-purpose pretraining and customizable fine-tuning.
Meta’s 7B-parameter instruction-tuned model optimized for chat, dialogue, and assistant-style applications.
Meta’s experimental ultra-sparse MoE model with 128 experts, designed to explore efficient large-scale scaling and routing strategies for future LLaMA architectures.
Meta’s experimental LLaMA 4-series MoE model with 17 billion parameters and 16 experts, designed to explore sparse routing and scaling strategies.
A next-generation 8-billion-parameter open-weight language model from Meta, optimized for reasoning and general-purpose tasks.
A powerful sparse Mixture-of-Experts (MoE) instruction-tuned language model by Mistral AI, combining efficiency and performance for chat and task-oriented generation.
A 30-billion-parameter open-source language model from MosaicML — a strong, general-purpose LLM balancing scale, performance, and inference efficiency.
An advanced, text-only language model from OpenAI with GPT-4-level capabilities, optimized for performance, efficiency, and competitive benchmarks.
A 32-billion-parameter large language model developed by Qwen LM, designed to deliver high-quality instruction following and multilingual chat capabilities.