JUHE API Marketplace
Text Generation

DeepSeek-V3

DeepSeek-V3.1 is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes via prompt templates.

Available on Wisdom Gate
Provider
deepseek
Parameters
671b
Context Length
131k
Release Date
2024-12-25
License
mit_+_model_license_(commercial_use_allowed)
Country
CN 🇨🇳
Model Introduction

DeepSeek-V3.1 is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes via prompt templates. It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference. Users can control the reasoning behaviour with the reasoning enabled boolean. Learn more in our docs

The model improves tool use, code generation, and reasoning efficiency, achieving performance comparable to DeepSeek-R1 on difficult benchmarks while responding more quickly. It supports structured tool calling, code agents, and search agents, making it suitable for research, coding, and agentic workflows.

It succeeds the DeepSeek V3-0324 model and performs well on a variety of tasks.

Specs

Provider
deepseek
Parameters
671b
Context Length
131k
Release Date
2024-12-25
License
mit_+_model_license_(commercial_use_allowed)
Country
CN 🇨🇳

Score

DROP
0.916
GPQA
0.591
SWE-bench Verified
0.420

Boost your DeepSeek-V3 with Wisdom Gate LLM API

Supporting GPT-5, Claude-4, DeepSeek v3, Gemini and more.

Enjoy a free trial and save 20%+ compared to official pricing.

DeepSeek-V3 Model: Specs, Pricing and Capabilities | JuheAPI