llama-4-scout by deepinfra - AI Model Details, Pricing, and Performance Metrics

meta
llama-4-scout
meta

llama-4-scout

completions
bydeepinfra

Llama 4 Scout 17B Instruct (16E) is a mixture-of-experts (MoE) language model developed by Meta, activating 17 billion parameters out of a total of 109B. It supports native multimodal input (text and image) and multilingual output (text and code) across 12 supported languages. Designed for assistant-style interaction and visual reasoning, Scout uses 16 experts per forward pass and features a context length of 10 million tokens, with a training corpus of ~40 trillion tokens. Built for high efficiency and local or commercial deployment, Llama 4 Scout incorporates early fusion for seamless modality integration. It is instruction-tuned for use in multilingual chat, captioning, and image understanding tasks. Released under the Llama 4 Community License, it was last trained on data up to August 2024 and launched publicly on April 5, 2025.

Released
Apr 5, 2025
Knowledge
Oct 7, 2024
License
llama_4_community_license_agreement
Context
327680
Input
$0.08 / 1M tokens
Output
$0.3 / 1M tokens
Accepts: text, image
Returns: text

Access llama-4-scout through LangDB AI Gateway

Recommended

Integrate with meta's llama-4-scout and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.

Unified API
Cost Optimization
Enterprise Security
Get Started Now

Free tier available • No credit card required

Instant Setup
99.9% Uptime
10,000+Monthly Requests

Category Scores

Benchmark Tests

View Other Benchmarks
AIME
28.3
Mathematics
AA Coding Index
23.5
Programming
AAII
28.1
General
AA Math Index
14.0
Mathematics
GPQA
57.9
STEM (Physics, Chemistry, Biology)
HLE
4.3
General Knowledge
LiveCodeBench
29.9
Programming
MATH-500
84.4
Mathematics
MMLU
79.6
General Knowledge
MMLU-Pro
74.8
General Knowledge
MMMU
69.4
General Knowledge
SciCode
17.0
Scientific

Code Examples

Integration samples and API usage