OpenAI's high-intelligence flagship model for complex, multi-step tasks. GPT-4o is cheaper and faster than GPT-4 Turbo.
OpenAI's affordable and intelligent small model for fast, lightweight tasks, is cheaper and more capable than GPT-3.5 Turbo.
OpenAI's previous high-intelligence model, optimized for chat but works well for traditional completions tasks.
OpenAI's cost-efficient reasoning model, excels at STEM, especially math and coding—nearly matching the performance of OpenAI o1 on evaluation benchmarks such as AIME and Codeforces.
The latest and strongest model family from OpenAI, o1 is designed to spend more time thinking before responding. The o1 model series is trained with large-scale reinforcement learning to reason using chain of thought. The o1 models are optimized for math, science, programming, and other STEM-related tasks. They consistently exhibit PhD-level accuracy on benchmarks in physics, chemistry, and biology.
Claude 3.5 Sonnet is a high-speed, cost-effective model offering industry-leading performance in reasoning, knowledge, and coding. It operates twice as fast as its predecessor. Key features include enhanced humor and nuance understanding, advanced coding capabilities, and strong visual reasoning.
Claude 3.5 Haiku features offers enhanced capabilities in speed, coding accuracy, and tool use. Engineered to excel in real-time applications, it delivers quick response times that are essential for dynamic tasks such as chat interactions and immediate coding suggestions.
Anthropic's Claude 3 Opus can handle complex analysis, longer tasks with multiple steps, and higher-order math and coding tasks, it provides top-level performance, intelligence, fluency, and understanding.
Anthropic's Claude 3 Haiku outperforms models in its intelligence category on performance, speed and cost without the need for specialized fine-tuning, it's the fastest and most compact model for near-instant responsiveness.
The Gemini 1.5 Pro is a cutting-edge multimodal AI model developed by Google DeepMind. It excels in processing and understanding text, images, audio, and video, featuring a breakthrough long context window of up to 1 million tokens. This model powers generative AI services across Google's platforms and supports third-party developers.
Gemini 1.5 Flash is a cutting-edge multimodal AI model known for its speed and efficiency. It excels in tasks like visual understanding and classification, featuring a long context window of up to one million tokens. This model is optimized for high-volume, high-speed applications, making it a significant advancement in AI technology.
Gemini 2.0 Flash builds on the success of 1.5 Flash, offering improved performance and twice the speed of 1.5 Pro on key benchmarks. It supports multimodal inputs like images, video, and audio, as well as outputs such as generated images, text, and multilingual text-to-speech. Additionally, it can natively integrate with tools like Google Search, execute code, and use third-party functions.
Gemini 2.0 Flash Thinking is an experimental AI model that showcases enhanced reasoning capabilities through transparent thought processes. The model demonstrates visible planning steps, tackles complex problems at unprecedented speeds, and offers expanded functionalities. This breakthrough innovation allows users to observe the AI's cognitive process in real-time while delivering swift, sophisticated solutions.
DeepSeek-V3 is the latest open-source model from DeepSeek, DeepSeek-V3 has outperformed other open-source models like Qwen2.5-72B and Llama-3.1-405B in various evaluations, and its performance is on par with world-class closed-source models like GPT-4o and Claude-3.5-Sonnet.
Grok-2 is xAI's frontier language model with state-of-the-art reasoning capabilities.
The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model in 70B (text in/text out). The Llama 3.3 instruction tuned text only model is optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.
Instruction-tuned image reasoning model with 90B parameters from Meta. Optimized for visual recognition, image reasoning, captioning, and answering general questions about an image. The model can understand visual data, such as charts and graphs and also bridge the gap between vision and language by generating text to describe images details.
The Meta Llama 3.1 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction tuned generative models in 8B, 70B and 405B sizes. The Llama 3.1 instruction tuned text only models (8B, 70B, 405B) are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks. 405B model is the most capable from the Llama 3.1 family.
Mistral Large is a cutting-edge language model developed by Mistral AI, renowned for its advanced reasoning capabilities. It excels in multilingual tasks, code generation, and complex problem-solving, making it ideal for diverse text-based applications.
Pixtral Large is a 124B open-weights multimodal model built on top of Mistral Large 2. The model is able to understand documents, charts and natural images.
Gemma 2 is a state-of-the-art, lightweight open model developed by Google, available in 9 billion and 27 billion parameter sizes. It offers enhanced performance and efficiency, building on the technology used in the Gemini models. Designed for a wide range of applications, Gemma 2 excels in text-to-text tasks, making it a versatile tool for developers.
The Perplexity Sonar Online model is a state-of-the-art large language model developed by Perplexity AI. It offers real-time internet access, ensuring up-to-date information retrieval. Known for its cost-efficiency, speed, and enhanced performance, it surpasses previous models in the Sonar family, making it ideal for dynamic and accurate data processing.
The new Command R+ model delivers roughly 50% higher throughput and 25% lower latencies as compared to the previous Command R+ version, while keeping the hardware footprint the same.
Qwen2.5 is a model pretrained on a large-scale dataset of up to 18 trillion tokens, offering significant improvements in knowledge, coding, mathematics, and instruction following compared to its predecessor Qwen2. The model also features enhanced capabilities in generating long texts, understanding structured data, and generating structured outputs, while supporting multilingual capabilities for over 29 languages.
QwQ is an experimental research model developed by the Qwen Team, designed to advance AI reasoning capabilities. This model embodies the spirit of philosophical inquiry, approaching problems with genuine wonder and doubt. QwQ demonstrates impressive analytical abilities, achieving scores of 65.2% on GPQA, 50.0% on AIME, 90.6% on MATH-500, and 50.0% on LiveCodeBench. With its contemplative approach and exceptional performance on complex problems.
The Yi Large model was designed by 01.AI with the following usecases in mind: knowledge search, data classification, human-like chat bots, and customer service. It stands out for its multilingual proficiency, particularly in Spanish, Chinese, Japanese, German, and French.
DBRX is a state-of-the-art, transformer-based, decoder-only large language model developed by Databricks. It features a Mixture-of-Experts (MoE) architecture with 132 billion parameters, designed for efficient next-token prediction. Released in 2024, it outperforms many open-source models on standard benchmarks.
WizardLM-2 8x22B is Microsoft AI's most advanced Wizard model. It demonstrates highly competitive performance compared to leading proprietary models, and it consistently outperforms all existing state-of-the-art opensource models.
Amazon Nova Pro 1.0 is a capable multimodal model from Amazon focused on providing a combination of accuracy, speed, and cost for a wide range of tasks. As of December 2024, it achieves state-of-the-art performance on key benchmarks including visual question answering (TextVQA) and video understanding (VATEX).