This is a snapshot of the popular LLMs that are available on the Hatz AI platform.
DISCLAIMER: The newest models from each developer tend to be the "best" option. Many LLMs return similar or comparable results for the majority of business use cases.
Don't worry about choosing the "right" model - try several of the newest ones and choose your favorite.
Newest models from each developer are marked in blue.
Cheap = lower credit consumption
LLM Name | Cost | Intelligence | Speed | Tools (Y/N) | Reasoning (Y/N) | What is it good at? |
Gemini 3 Pro Image (Nano Banana Pro) | Premium | High | Medium | N | N | best image generation |
Gemini 3 Pro Preview | Medium Low Cost | High | Medium | Y | Y | newest Gemini (testing) |
Claude 4.5 Sonnet | Medium High Cost | High | Medium | Y | Y | strong all‑purpose |
Claude Haiku 4.5 | Medium Low Cost | Medium | Fast | Y | N | cheapest fast agents |
GPT-5.2 | Medium High Cost | High | Medium | Y | Y | top reasoning + coding |
Gemini 2.5 Flash | Lowest Cost | Medium | Fast | Y | N | fast production work |
Amazon Nova 2 Lite | Low Cost | Medium | Fast | Y | Y | cheaper, fast general chat |
Amazon Nova Lite | Lowest Cost | Medium | Fast | Y | N | fast summaries + chat |
Amazon Nova Micro | Lowest Cost | Low | Fast | Y | Y | fast text tasks, lower credit usage |
Amazon Nova Pro | Low Cost | High | Medium | Y | N | reasoning + coding |
Claude 3 Haiku | Low Cost | Medium | Fast | Y | N | very fast summaries |
Claude 3 Sonnet | Medium High Cost | High | Medium | Y | N | writing + analysis |
Claude 3.5 Haiku | Low Cost | Medium | Fast | N | N | fast responses |
Claude 3.5 Sonnet v2 | Lowest Cost | High | Medium | Y | N | writing + coding |
Claude 3.7 Sonnet | Medium High Cost | High | Medium | Y | N | writing + coding |
Claude 4 Opus | Premium | High | Medium | Y | Y | reasoning + coding |
Claude 4 Sonnet | Medium High Cost | High | Medium | Y | Y | agents + coding |
Claude 4.5 Opus | Medium High Cost | High | Medium | Y | Y | reasoning + agents |
Deepseek R1 | Medium Low Cost | High | Slow | N | Y | hard problem solving |
Gemini 1.5 Flash | Low Cost | Medium | Fast | N | N | fast long-doc summaries |
Gemini 1.5 Pro | Low Cost | High | Medium | N | N | deep long-doc analysis |
Gemini 2.0 Flash | Lowest Cost | Medium | Fast | N | N | fast general chat |
Gemini 2.0 Flash Lite | Lowest Cost | High | Fast | N | N | cheapest bulk text |
Gemini 2.5 Flash Image (Nano Banana) | Low Cost | Medium | Fast | N | N | fast image generation |
Gemini 2.5 Pro | Medium High Cost | High | Fast | Y | Y | best Gemini reasoning |
Llama 3.1 70B Instruct | Medium Low Cost | Medium | Fast | N | N | quality open chat |
Llama 3.1 8B Instruct | Lowest Cost | Medium | Medium | N | N | cheap open chat |
Llama 3.2 11B Instruct | Medium Low Cost | High | Slow | N | N | balanced open chat |
Llama 3.2 1B Instruct | Medium Low Cost | Medium | Fast | N | N | ultra-cheap, tiny tasks |
Llama 4 Maverick 17B Instruct | Lowest Cost | High | Fast | Y | N | stronger mid-size chat |
Llama 4 Scout 17B Instruct | High Cost | High | Fast | Y | N | fast drafting + agents |
Mistral 7B Instruct | Lowest Cost | Low | Medium | N | N | cheap basic chat |
Mistral 8x7B Instruct | Lowest Cost | Medium | Fast | N | N | efficient bulk summaries |
Mistral Large | Medium High Cost | High | Medium | N | N | high-quality writing |
Mistral Large 3 | Medium High Cost | High | Medium | N | N | best Mistral reasoning |
GPT 3.5 Turbo | Low Cost | Medium | Fast | Y | N | cheap basic chat |
GPT 4 | High Cost | High | Medium | Y | N | strong general |
GPT 4.5 Preview | Premium | High | Medium | Y | N | new (testing) |
GPT 4o | Medium Low Cost | High | Medium | Y | N | vision + chat |
GPT-4.1 | Medium Low Cost | High | Medium | Y | N | coding + analysis |
GPT-4.1 Mini | Low Cost | Medium | Fast | Y | N | cheap general assistant |
GPT-4.1 Nano | Lowest Cost | Medium | Fast | Y | N | ultra-cheap bulk text |
GPT-5 | Medium Low Cost | High | Medium | Y | Y | strong general + coding |
GPT-5 Chat | Medium Low Cost | High | Medium | Y | N | chat assistant |
GPT-5 Mini | Low Cost | Medium | Fast | Y | Y | cheap agents + summaries |
GPT-5 Nano | Lowest Cost | Medium | Fast | Y | Y | fastest, cheap text |
GPT-5.1 | Medium Low Cost | High | Medium | Y | Y | strong quality general |
o1 | High Cost | High | Slow | Y | Y | reasoning |
o3 | High Cost | High | Slow | Y | Y | strong reasoning |
o3 mini | Medium Low Cost | Medium | Medium | Y | Y | cheaper reasoning |
o4 mini | Medium Low Cost | High | Medium | Y | Y | fast cheap reasoning |
Grok 2 | Medium Low Cost | Medium | Medium | Y | N | general chat |
Grok 2 Vision | Medium Low Cost | Medium | Medium | Y | N | vision + OCR |
Grok 3 | Medium High Cost | High | Medium | Y | N | reasoning + agents |
Grok 3 Fast | Medium High Cost | Medium | Fast | Y | N | fastest chat |
Grok 3 Mini | Lowest Cost | Medium | Fast | Y | N | cheap general |
Grok 3 Mini Fast | High Cost | Medium | Fast | Y | N | cheap + fast |
Grok 4 | Medium High Cost | High | Medium | Y | Y | research + reasoning |
Some LLM Terminology:
Token
The smallest unit of data that an LLM processes
Usually about 3/4 of a word
If you want to see how many INPUT tokens something is, use this site (it works best for OpenAI models):
***In some cases, tokens are not just words. If you upload a PDF with images, tokens will applied to interpret the images in addition to the words. If you upload a PNG or JPG with no words, tokens will still be used to understand the image.
Context Window
The segment of text that an AI model considers at any given moment when processing or generating language
You can think of this as the size of the input
If you upload a file (PDF, PNG, XLSX, etc) that is below the 30MB limit, but it is larger than the context window for the selected LLM, the request will fail. Try switching to a larger model, like Gemini
Knowledge Cutoff Date
The date that the LLM stopped "learning" new information. If you ask GPT 3.5 Turbo about an event that happened yesterday, it will not have that data
Reasoning
Reasoning models are a type of AI model that are trained to engage in "thinking" before providing an answer, breaking down problems into steps and showing their reasoning process, unlike models that directly generate answers.
They can take more time to answer because they are reflecting on their analysis step-by-step.
Vision
Computer vision gives LLMs the ability to "see" and interpret visual data like images and videos, enabling them to identify objects, classify scenes, and perform tasks that mimic human vision.
More Info
To dive deeper into these LLMs, check out these resources*
*Some features outlined on these sites might not be available on the Hatz AI platform, or they might work differently than explained here. Use these articles to better understand the token limits, reasoning models, and use cases.
