llama-4-scout by deepinfra - AI Model Details, Pricing, and Performance Metrics

meta
llama-4-scout
Try
meta

llama-4-scout

completions
bydeepinfra

Llama 4 Scout 17B Instruct (16E) is a mixture-of-experts (MoE) language model developed by Meta, activating 17 billion parameters out of a total of 109B. It supports native multimodal input (text and image) and multilingual output (text and code) across 12 supported languages. Designed for assistant-style interaction and visual reasoning, Scout uses 16 experts per forward pass and features a context length of 10 million tokens, with a training corpus of ~40 trillion tokens. Built for high efficiency and local or commercial deployment, Llama 4 Scout incorporates early fusion for seamless modality integration. It is instruction-tuned for use in multilingual chat, captioning, and image understanding tasks. Released under the Llama 4 Community License, it was last trained on data up to August 2024 and launched publicly on April 5, 2025.

Released
Apr 5, 2025
Knowledge
Oct 7, 2024
License
llama_4_community_license_agreement
Context
327680
Input
$0.08 / 1M tokens
Output
$0.3 / 1M tokens
Accepts: text, image
Returns: text

Access llama-4-scout through LangDB AI Gateway

Recommended

Integrate with meta's llama-4-scout and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.

Unified API
Cost Optimization
Enterprise Security
Get Started Now

Free tier available • No credit card required

Instant Setup
99.9% Uptime
10,000+Monthly Requests

Category Scores

Benchmark Tests

View Other Benchmarks
HLE
4.3
General Knowledge
AIME
28.3
Mathematics
GPQA
57.9
STEM (Physics, Chemistry, Biology)
MMMU
69.4
General Knowledge
SciCode
17.0
Scientific
MATH-500
84.4
Mathematics
MMLU-Pro
75.2
General Knowledge
LiveCodeBench
29.9
Programming
AA Math Index
14.0
Mathematics
AA Coding Index
6.7
Programming
AAII
13.5
General

Code Examples

Integration samples and API usage