chat models
Function Calling
deepseek-r1-70b
DeepSeek-R1-Distill-Llama-70B is a 70B-parameter distilled LLM, combining reasoning, speed, and accuracy for code, math, and complex logic tasks.
Price per million tokens
Input: 0.6€ / Output: 2.7€Vision
gemma-3-27b-it
Gemma 3 is a lightweight, multimodal open model by Google with 128K context, multilingual support, and broad deployment versatility.
Price per million tokens
Input: 0.95€ / Output: 5.5€Function Calling
gpt-oss-120b
GPT-OSS-120B is an open-weight 117B-parameter Mixture-of-Experts model by OpenAI, using only 5.1B active parameters per token. Supports reasoning, chain-of-thought, tool use, and fine-tuning.
Price per million tokens
Input: 1€ / Output: 4.2€Llama-3.1-8B-Instruct
LLaMA 3.1 8B is a multilingual model optimized for dialogue, outperforming many open and closed-source chat models.
Price per million tokens
Input: 0.05€ / Output: 0.25€Function Calling
Llama-3.3-70B-Instruct
Meta's multilingual 70B parameter language model, optimized for instruction-based dialogue and benchmarking.
Price per million tokens
Input: 0.6€ / Output: 2.7€llama-guard3-8b
Llama-Guard-3-8B is a fine-tuned Llama 3.1-8B model for multilingual content safety classification, detecting unsafe prompts and responses using MLCommons hazard taxonomy.
Price per million tokens
Input: 0.09€ / Output: 0.9€maestrale-chat-v0.4-beta
Mistral-7b for the Italian language, continued pre-training for Italian on a curated large-scale high-quality corpus. Powered by mii-llm
Price per million tokens
Input: 0.05€ / Output: 0.25€Function Calling
mistral-small3.2
Mistral-Small-3.2-24B-Instruct-2506: 24B multimodal instruction-tuned model, optimized for reasoning and STEM, supports robust function calling, reduces repetition, handles both text and vision inputs efficiently.
Price per million tokens
Input: 0.5€ / Output: 2.2€Phi-4
Microsoft's phi-4 combines academic and synthetic datasets to enhance reasoning abilities.
Price per million tokens
Input: 0.05€ / Output: 0.25€Vision
Qwen2.5-VL-32B-Instruct
Qwen understands images, videos, and text, reasons deeply, generates structured outputs, localizes objects accurately, acts agentically, fast, reliable, flexible, smart.
Price per million tokens
Input: 0.45€ / Output: 2.5€Qwen3-8B
Qwen3 is the latest generation of large language models in Qwen series.
Price per million tokens
Input: 0.07€ / Output: 0.35€Function Calling
qwen3-coder-30b
A 30 B‑parameter MoE coding model, optimized for agentic coding with long 256 K token context, supports function‑call format and strong instruction‑based reasoning
Price per million tokens
Input: 0.5€ / Output: 2€