LLM token cost

Is asking a LLM for a short answer to a question a cost effective idea?

Published: Thursday, Apr 4, 2024 Last modified: Monday, May 27, 2024

Mistral

https://docs.mistral.ai/platform/pricing/

Mistral Large

Input tokens

Output tokens

Prompt limit: 2048 tokens

OpenAI

https://openai.com/pricing

GPT-4 Turbo

Context tokens

Generated tokens

Prompt limit: 2048 tokens

Limits

https://help.openai.com/en/articles/4936856-what-are-tokens-and-how-to-count-them

Depending on the model used, requests can use up to 128,000 tokens shared between prompt and completion. Some models, like GPT-4 Turbo, have different limits on input and output tokens. There are often creative ways to solve problems within the limit, e.g. condensing your prompt, breaking the text into smaller pieces, etc.

Anthropic

Claude 3 - Opus