claude-3.7-sonnet
completions
Intelligent model, with visible step‑by‑step reasoning
Input:$3 / 1M tokens•$0.3 / 1M tokenscached read•$3.75 / 1M tokenscached write
Output:$15 / 1M tokens
Context:200K tokens
tools
text
text
image
Access claude-3.7-sonnet through LangDB AI Gateway
Recommended
Integrate with anthropic's claude-3.7-sonnet and 250+ other models through a unified API. Monitor usage, control costs, and enhance security.
Unified API
Cost Optimization
Enterprise Security
Get Started Now
Free tier available • No credit card required
Instant Setup
99.9% Uptime
10,000+Monthly Requests
Code Example
Configuration
Base URL
API Keys
Headers
Project ID in header
X-Run-Id
X-Thread-Id
Model Parameters
6 availableinclude_reasoning
max_tokens
stop
temperature
012
tool_choice
tools
Additional Configuration
Tools
Guards
User:
Id:
Name:
Tags:
Publicly Shared Threads5
- Discussion about The Impact of AI on Social Media Marketing in 2024
- 3 months agoPython function using a hash map to find indices of two numbers in a list that sum to a target with O(n) time complexity.two sum problempython two sum functiontwo sum algorithmtwo sum hash map
- 3 months agoCreated a CSS-only image slider toggling three images via radio inputs and styled label arrows with smooth fade transitions.css image sliderhtml css sliderpure css image carouselimage slider without javascript
- 3 months agoResponsive navigation bar using HTML, CSS, and JavaScript that switches to a hamburger menu and toggles links on screens below 600px.responsive navigation barhamburger menucss media queriesvanilla javascript toggle
Popular Models10
- deepseek-chat-v3-0324deepseekDeepSeek V3, a 685B-parameter, mixture-of-experts model, is the latest iteration of the flagship chat model family from the DeepSeek team. It succeeds the [DeepSeek V3](/deepseek/deepseek-chat-v3) model and performs really well on a variety of tasks.Input:$0.24 / 1M tokensOutput:$0.82 / 1M tokensContext:163840 tokenstoolstexttext
- gpt-4o-miniopenaiGPT-4o mini (o for omni) is a fast, affordable small model for focused tasks. It accepts both text and image inputs, and produces text outputs (including Structured Outputs). It is ideal for fine-tuning, and model outputs from a larger model like GPT-4o can be distilled to GPT-4o-mini to produce similar results at lower cost and latency.The knowledge cutoff for GPT-4o-mini models is October, 2023.Input:$0.15 / 1M tokensOutput:$0.6 / 1M tokensContext:128K tokenstoolstextimagetext
- Qwen3-235B-A22B-Instruct-2507 is a multilingual, instruction-tuned mixture-of-experts language model based on the Qwen3-235B architecture, with 22B active parameters per forward pass. It is optimized for general-purpose text generation, including instruction following, logical reasoning, math, code, and tool usage. The model supports a native 262K context length and does not implement "thinking mode" (<think> blocks). Compared to its base variant, this version delivers significant gains in knowledge coverage, long-context reasoning, coding benchmarks, and alignment with open-ended tasks. It is particularly strong on multilingual understanding, math reasoning (e.g., AIME, HMMT), and alignment evaluations like Arena-Hard and WritingBench.Input:$0.13 / 1M tokensOutput:$0.66 / 1M tokensContext:262144 tokenstexttext
- llama-3.2-1b-instructmeta-llamaLlama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance. Supporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models. Click here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md). Usage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).Input:$0.01 / 1M tokensOutput:$0.07 / 1M tokensContext:131072 tokenstexttext
- deepseek-r1-0528deepseekMay 28th update to the [original DeepSeek R1](/deepseek/deepseek-r1) Performance on par with [OpenAI o1](/openai/o1), but open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass. Fully open-source model.Input:$0.39 / 1M tokensOutput:$1.63 / 1M tokensContext:163840 tokenstexttext
- qwen3-235b-a22bqwenQwen3-235B-A22B is a 235B parameter mixture-of-experts (MoE) model developed by Qwen, activating 22B parameters per forward pass. It supports seamless switching between a "thinking" mode for complex reasoning, math, and code tasks, and a "non-thinking" mode for general conversational efficiency. The model demonstrates strong reasoning ability, multilingual support (100+ languages and dialects), advanced instruction-following, and agent tool-calling capabilities. It natively handles a 32K token context window and extends up to 131K tokens using YaRN-based scaling.Input:$0.18 / 1M tokensOutput:$0.6 / 1M tokensContext:40960 tokenstoolstexttext
- deepseek-chatdeepseekDeepSeek-Chat is an advanced conversational AI model designed to provide intelligentInput:$0.27 / 1M tokensOutput:$1.1 / 1M tokensContext:128K tokenstoolstexttext
- deepseek-v3-basedeepseekNote that this is a base model mostly meant for testing, you need to provide detailed prompts for the model to return useful responses. DeepSeek-V3 Base is a 671B parameter open Mixture-of-Experts (MoE) language model with 37B active parameters per forward pass and a context length of 128K tokens. Trained on 14.8T tokens using FP8 mixed precision, it achieves high training efficiency and stability, with strong performance across language, reasoning, math, and coding tasks. DeepSeek-V3 Base is the pre-trained model behind [DeepSeek V3](/deepseek/deepseek-chat-v3)Input:$0.2 / 1M tokensOutput:$0.8 / 1M tokensContext:163840 tokenstexttext
- glm-4.5-airz-aiGLM-4.5-Air is the lightweight variant of our latest flagship model family, also purpose-built for agent-centric applications. Like GLM-4.5, it adopts the Mixture-of-Experts (MoE) architecture but with a more compact parameter size. GLM-4.5-Air also supports hybrid inference modes, offering a "thinking mode" for advanced reasoning and tool use, and a "non-thinking mode" for real-time interaction. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config)Input:$0.2 / 1M tokensOutput:$1.1 / 1M tokensContext:131072 tokenstoolstexttext
- claude-sonnet-4anthropicOur high-performance model with exceptional reasoning and efficiencyInput:$3 / 1M tokensOutput:$15 / 1M tokensContext:200K tokenstoolstextimagetext
Related AI Model Resources
Explore more AI models, providers, and integration options: