{
    "chat": {
        "open-mistral-7b": {
            "description": "Our very first. A 7B transformer model, fast-deployed and easily customisable. Small, yet very powerful for a variety of use cases. English and code.",
            "inputTokens": 32000,
            "inputTokenPrice": 0.25,
            "outputTokens": 8191,
            "outputTokenPrice": 0.25
        },
        "open-mixtral-8x7b": {
            "description": "A 7B sparse Mixture-of-Experts (SMoE). Uses 12B active parameters out of 45B total. Fluent in English, French, Italian, German, Spanish, and strong in code.",
            "inputTokens": 32000,
            "inputTokenPrice": 0.7,
            "outputTokens": 8191,
            "outputTokenPrice": 0.7
        },
        "open-mixtral-8x22b": {
            "description": "A 22B sparse Mixture-of-Experts (SMoE). Uses only 39B active parameters out of 141B. Fluent in English, French, Italian, German, Spanish, and strong in code.",
            "inputTokens": 65336,
            "inputTokenPrice": 2,
            "outputTokens": 8191,
            "outputTokenPrice": 6
        },
        "mistral-small-latest": {
            "description": "Cost-efficient reasoning for low-latency workloads. Fluent in English, French, Italian, German, Spanish, and strong in code.",
            "inputTokens": 131072,
            "inputTokenPrice": 0.06,
            "outputTokens": 131072,
            "outputTokenPrice": 0.18
        },
        "mistral-medium-latest": {
            "description": "Balanced reasoning for a wide range of tasks. Fluent in English, French, Italian, German, Spanish, and strong in code.",
            "inputTokens": 131072,
            "inputTokenPrice": 0.4,
            "outputTokens": 131072,
            "outputTokenPrice": 2
        },
        "mistral-large-latest": {
            "description": "Top-tier reasoning for high-complexity tasks. Fluent in English, French, Italian, German, Spanish, and strong in code.",
            "inputTokens": 262144,
            "inputTokenPrice": 0.5,
            "outputTokens": 262144,
            "outputTokenPrice": 1.5
        },
        "mistral-tiny": {
            "description": "",
            "inputTokens": 32000,
            "inputTokenPrice": 0.25,
            "outputTokens": 8191,
            "outputTokenPrice": 0.25
        },
        "mistral-small": {
            "description": "",
            "inputTokens": 32000,
            "inputTokenPrice": 0.1,
            "outputTokens": 8191,
            "outputTokenPrice": 0.3
        },
        "mistral-medium": {
            "description": "",
            "inputTokens": 32000,
            "inputTokenPrice": 2.7,
            "outputTokens": 8191,
            "outputTokenPrice": 8.1
        },
        "codestral-latest": {
            "description": "",
            "inputTokens": 32000,
            "inputTokenPrice": 1,
            "outputTokens": 8191,
            "outputTokenPrice": 3
        },
        "open-mistral-nemo": {
            "description": "https:\/\/mistral.ai\/technology\/",
            "inputTokens": 128000,
            "inputTokenPrice": 0.3,
            "outputTokens": 128000,
            "outputTokenPrice": 0.3
        },
        "open-codestral-mamba": {
            "description": "https:\/\/mistral.ai\/technology\/",
            "inputTokens": 256000,
            "inputTokenPrice": 0.25,
            "outputTokens": 256000,
            "outputTokenPrice": 0.25
        },
        "codestral-mamba-latest": {
            "description": "https:\/\/mistral.ai\/technology\/",
            "inputTokens": 256000,
            "inputTokenPrice": 0.25,
            "outputTokens": 256000,
            "outputTokenPrice": 0.25
        },
        "pixtral-large-latest": {
            "description": "",
            "inputTokens": 128000,
            "inputTokenPrice": 2,
            "outputTokens": 128000,
            "outputTokenPrice": 6
        },
        "devstral-small-latest": {
            "description": "https:\/\/docs.mistral.ai\/models\/devstral-small-2-25-12",
            "inputTokens": 256000,
            "inputTokenPrice": 0.1,
            "outputTokens": 256000,
            "outputTokenPrice": 0.3
        },
        "devstral-latest": {
            "description": "https:\/\/mistral.ai\/news\/devstral-2-vibe-cli",
            "inputTokens": 256000,
            "inputTokenPrice": 0.4,
            "outputTokens": 256000,
            "outputTokenPrice": 2
        },
        "devstral-medium-latest": {
            "description": "https:\/\/mistral.ai\/news\/devstral-2-vibe-cli",
            "inputTokens": 256000,
            "inputTokenPrice": 0.4,
            "outputTokens": 256000,
            "outputTokenPrice": 2
        },
        "magistral-medium-latest": {
            "description": "https:\/\/mistral.ai\/news\/magistral",
            "inputTokens": 40000,
            "inputTokenPrice": 2,
            "outputTokens": 40000,
            "outputTokenPrice": 5
        },
        "magistral-small-latest": {
            "description": "https:\/\/mistral.ai\/pricing#api-pricing",
            "inputTokens": 40000,
            "inputTokenPrice": 0.5,
            "outputTokens": 40000,
            "outputTokenPrice": 1.5
        },
        "mistral-large-3": {
            "description": "https:\/\/docs.mistral.ai\/models\/mistral-large-3-25-12",
            "inputTokens": 262144,
            "inputTokenPrice": 0.5,
            "outputTokens": 262144,
            "outputTokenPrice": 1.5
        }
    },
    "embedding": {
        "mistral-embed": {
            "description": "State-of-the-art semantic for extracting representation of text extracts. English only for now.",
            "inputTokens": 8192,
            "inputTokenPrice": 0.1,
            "dimensions": 1024
        },
        "codestral-embed": {
            "description": "",
            "inputTokens": 8192,
            "inputTokenPrice": 0.15,
            "dimensions": 1536
        }
    }
}