Skip to main content

Explore Mistral Models & APIs with APIpie Overview

Mistral AI
info

For detailed information about using models with APIpie, check out our Models Overview and Completions Guide.

Description

Mistral AI, a leading European AI company founded by former DeepMind and Meta AI researchers, develops powerful large language models known for their efficiency and performance. Their models, including the groundbreaking Mixtral-8x7B (which outperforms GPT-3.5)(one of the best 7B models), consistently achieve top rankings in open-source model benchmarks. The Mixtral architecture uses a Mixture of Experts approach, allowing it to achieve remarkable performance while maintaining computational efficiency. These models are available through various providers integrated with APIpie's routing system.

Key Features

  • Extended Token Capacity: Models support context lengths from 4K to 256K tokens, with specialized variants like Codestral-Mamba (256K tokens) and Mistral-Nemo (128K tokens).
  • Multi-Provider Availability: Accessible through OpenRouter, Together, Deepinfra, Bedrock, EdenAI.
  • Diverse Applications: Optimized for chat, instruction-following, code generation, and embeddings, with specialized models for different tasks.
  • Efficiency: High performance with optimized parameter counts and innovative architectures like Mixture of Experts and Mamba.

Model List in the Mistral Series

Model List updates dynamically please see the Models Route for the up to date list of models

info

For information about model performance and benchmarks, see the Mistral AI Documentation and Model Cards.

Model NameMax TokensResponse TokensProvidersSubtype
Mixtral Models
Mixtral-8x22B-Instruct-v0.165,53665,536Deepinfra, Together, OpenRouterChat
Mixtral-8x7B-Instruct-v0.132,76832,768Deepinfra, TogetherChat
Mixtral-8x7B-v0.132,76832,768TogetherLanguage
mixtral-8x7b-instruct32,76832,768OpenRouterChat
mixtral-8x7b32,76832,768OpenRouterChat
mixtral-8x7b-instruct-v04,0964,096BedrockChat
Mistral Models
Mistral-7B-Instruct-v0.14,0964,096Deepinfra, Together, OpenRouterChat
Mistral-7B-Instruct-v0.232,76832,768Deepinfra, Together, MonsterChat
Mistral-7B-Instruct-v0.332,7684,096Together, OpenRouterChat
Mistral-7B-v0.14,0964,096TogetherLanguage
mistral-7b-instruct32,76832,768OpenRouterChat
mistral-7b-instruct-v08,1928,192BedrockChat
mistral-large128,000128,000OpenRouterChat
mistral-large8,1928,192BedrockChat
mistral-medium32,00032,000OpenRouterChat
mistral-small32,00032,000OpenRouterChat
mistral-tiny32,00032,000OpenRouterChat
mistral-small-2402-v18,1928,192BedrockChat
mistral-nemo128,0004,096OpenRouterChat
Specialized Models
codestral-mamba256,000256,000OpenRouterCode
openhermes-2.5-mistral-7b4,0964,096OpenRouterChat
pixtral-12b4,0964,096OpenRouterChat
pixtral-large-2411128,000128,000OpenRouterChat
ministral-8b128,000128,000OpenRouterChat
ministral-3b128,000128,000OpenRouterChat
Other Variants
1024__mistral-embed--EdenAIEmbedding
large-latest--EdenAIChat
tiny--EdenAIChat
small--EdenAIChat
ministral-3b-latest--EdenAIChat
ministral-8b-latest--EdenAIChat

Example API Call

Below is an example of how to use the Chat Completions API to interact with a Mistral model:

curl -L -X POST 'https://apipie.ai/v1/chat/completions' \
-H 'Content-Type: application/json' \
-H 'Accept: application/json' \
-H 'Authorization: Bearer <YOUR_API_KEY>' \
--data-raw '{
"provider": "openrouter",
"model": "mixtral-8x7b-instruct",
"max_tokens": 150,
"messages": [
{
"role": "user",
"content": "What are the key differences between traditional and quantum computing?"
}
]
}'

Response Example

{
"id": "chatcmpl-12345example12345",
"object": "chat.completion",
"created": 1729535643,
"provider": "openrouter",
"model": "mixtral-8x7b-instruct",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "Here are the key differences between traditional and quantum computing:\n\n1. **Information Processing**\n- Traditional: Uses bits (0 or 1)\n- Quantum: Uses qubits (can be 0, 1, or both simultaneously)\n\n2. **Computation Method**\n- Traditional: Sequential processing\n- Quantum: Parallel processing through superposition\n\n3. **Processing Power**\n- Traditional: Linear scaling\n- Quantum: Exponential scaling for certain tasks\n\n4. **Applications**\n- Traditional: General-purpose computing\n- Quantum: Specialized tasks like cryptography and complex simulations"
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 18,
"completion_tokens": 132,
"total_tokens": 150
}
}

API Highlights

  • Provider Selection: Choose from multiple providers or use automatic routing
  • Model Options: Select from various Mistral models based on your needs
  • Token Management: Set appropriate token limits based on model capabilities
  • Message Formatting: Structure your prompts effectively using message formatting

Applications and Integrations

  • Conversational AI: Build chatbots and virtual assistants.
  • Code Generation: Leverage specialized models like Codestral-Mamba for programming tasks
  • Enterprise Solutions: Deploy models for business applications
  • Extended Context Processing: Handle long documents with models supporting up to 256K tokens
  • Research Applications: Utilize open-weights models for academic research and development

Ethical Considerations

Mistral models should be used responsibly with appropriate safeguards. For guidance on responsible AI usage, see Mistral's terms.


Licensing

Mistral models are available under various licensing terms, including the Apache 2.0 License for open-source models. For detailed information, consult Mistral's documentation.

tip

Try out Mistral models in APIpie's supported integrations.