Qwen 3 14B
Qwen 3 model with 14B parameters offering excellent performance-to-size efficiency
Read more→Qwen 3 model with 14B parameters offering excellent performance-to-size efficiency
Read more→MoE Qwen 3 model with 30B total parameters, activating 3B for efficient inference
Read more→Large MoE Qwen 3 thinking model with 235B parameters and advanced reasoning capabilities
Read more→Qwen API model with 1M token context support for extensive document processing
Read more→AWS Nova lightweight model with 300K context for fast, cost-efficient text processing
Read more→AWS Nova advanced model with 300K context for complex reasoning and multimodal tasks
Read more→Advanced Haiku model matching Claude 3 Opus performance with maintained speed and cost-efficiency
Read more→Latest generation Sonnet with best-in-class performance for complex agents and coding tasks
Read more→Exceptional reasoning model for specialized complex tasks requiring advanced analytical capabilities
Read more→Hybrid reasoning model with extended thinking mode for complex problem-solving and quick responses
Read more→Updated Haiku model from October 2024 with enhanced accuracy and performance
Read more→DeepSeek MoE model with 671B parameters (37B active) and 128K context for general tasks
Read more→DeepSeek reasoning model from May 2024 with chain-of-thought capabilities for complex problems
Read more→Advanced DeepSeek reasoning model with RL training, comparable to OpenAI o1 in performance
Read more→Efficient MoE model with 671B parameters trained with FP8, achieving strong benchmark results
Read more→Advanced thinking model with adaptive reasoning, excellent price-performance, and multimodal capabilities
Read more→Optimized Gemini model for cost-efficiency, high throughput, and low latency applications
Read more→Most intelligent Gemini model with enhanced reasoning for complex coding, math, and scientific tasks
Read more→Open-source instruction-tuned Gemma model with 27B parameters for diverse language tasks
Read more→Specific Gemini 2.0 Flash version with stable performance and consistent behavior
Read more→Preview version of Gemini 2.5 Pro with advanced reasoning capabilities released in March 2025
Read more→Next-generation Llama 4 model with advanced capabilities and improved training efficiency
Read more→Llama 4 Scout variant with 17B parameters and mixture-of-experts architecture for efficiency
Read more→Llama 3.3 model with 70B parameters offering improved performance over 3.1 version
Read more→FP8-quantized 17B Llama 4 Maverick model optimized for deployment efficiency and speed
Read more→Latest Phi model with 14B parameters offering improved performance in reasoning and coding
Read more→Updated Mistral Large from November 2024 with improved performance and capabilities
Read more→Specialized Mistral model variant designed for specific enterprise use cases
Read more→Compact 24B parameter Mistral model optimized for cost-effective instruction following
Read more→Advanced Kimi model with improved reasoning depth and instruction following capabilities
Read more→NVIDIA-tuned 253B Llama 3.1 model optimized for enterprise applications and instruction following
Read more→NVIDIA optimized 49B Llama 3.3 model providing excellent performance-to-size ratio
Read more→NVIDIA tuned 70B Llama 3.1 model with enhanced instruction following and helpfulness
Read more→Smaller, faster, and more affordable version of GPT-4o, ideal for high-volume applications requiring good intelligence
Read more→Enhanced iteration of GPT-4 with improved reasoning, coding, and multimodal capabilities
Read more→Most advanced OpenAI reasoning model with multimodal capabilities and agentic tool use for complex analysis
Read more→Lightweight reasoning model balancing speed and intelligence for everyday complex tasks
Read more→Next-generation GPT model with enhanced reasoning, larger context window, and improved general capabilities
Read more→Efficient version of GPT-5 designed for high-throughput applications with cost optimization
Read more→Ultra-compact GPT-5 variant for edge deployment and resource-constrained environments
Read more→Open-source large language model with 120B parameters offering competitive performance
Read more→Compact GPT-4.1 variant optimized for efficiency while maintaining strong performance
Read more→January 2025 release of o3-mini with enhanced STEM capabilities and developer features
Read more→April 2025 o4-mini release with improved reasoning efficiency and balanced performance
Read more→Compact Grok model optimized for efficient deployment with maintained intelligence
Read more→Latest Grok model with advanced reasoning capabilities and extended thinking mode
Read more→Grok 2 version from December 2024 with incremental improvements and optimizations
Read more→Beta version of Grok 3 with extended reasoning for complex problem-solving tasks
Read more→Enhanced GLM model with improved agentic capabilities, reasoning, and coding performance
Read more→Lightweight GLM 4.5 variant optimized for faster inference and lower computational costs
Read more→