DeepSeek’s 685 billion parameter model is competing with Claude 3.7

DeepSeek’s 685 billion parameter model is competing with Claude 3.7
Source: Getty Image via Unsplash

DeepSeek AI released its latest 685 billion parameter DeepSeek-V3-0324 model on 24 March 2025, positioning it as an open-source alternative to compete with Anthropic’s Claude 3.7 Sonnet model. The new model demonstrates significant advancements in coding, mathematical tasks, and general problem-solving, while being freely available under an MIT licence.

The DeepSeek-V3-0324 employs a Mixture-of-Experts (MoE) architecture, activating only 37 billion parameters per token, enabling efficient operation even with limited resources. Its benchmark results are impressive: on the MMLU-Pro test, which measures multi-subject language understanding, performance improved from 75.9% to 81.2% (+5.3 points); on the GPQA test, assessing complex general knowledge, it rose from 59.1% to 68.4% (+9.3 points); and on the AIME test, evaluating medical context interpretation, it jumped from 39.6% to 59.4% (+19.8 points). In the LiveCodeBench test, which measures real-world programming tasks, the score increased from 39.2% to 49.2%, marking a 10-point improvement.

Source: https://huggingface.co/deepseek-ai/DeepSeek-V3-0324

The DeepSeek-V3-0324 is widely applicable across various industries, including the financial sector (complex analytics and risk assessment), healthcare (supporting medical research and diagnostic tools), software development (automated code generation and error analysis), and telecommunications (optimising network architectures). The model is accessible through various frameworks such as SGLang (for NVIDIA/AMD GPUs), LMDeploy, and TensorRT-LLM, with quantised versions in 1.78–4.5-bit GGUF formats also released, enabling local use even on less powerful hardware.

Sources:

1.

deepseek-ai/DeepSeek-V3-0324 · Hugging Face
We’re on a journey to advance and democratize artificial intelligence through open source and open science.

2.

DeepSeek-V3–0324: The most powerful open source AI model with 685 billion parameters
The new DeepSeek-V3–0324 represents a significant advance in the field of open source artificial intelligence. With a total of 685 billion parameters, this language assistant significantly…

3.

DeepSeek V3 0324 Impresses: Is It as Good as Claude 3.7 Sonnet? - Bind AI
DeepSeek V3 0324 impresses with strong coding performance, rivaling Claude 3.7 Sonnet. Read our analysis to see if it measures up in key AI tasks.

4.

DeepSeek V3-0324 vs Claude 3.7: Which is the Better Coder?
This DeepSeek V3-0324 vs Claude 3.7 comparison looks into the performance of both these models across benchmarks and coding tasks.