Update README.md
Browse files
README.md
CHANGED
|
@@ -24,9 +24,11 @@ tags:
|
|
| 24 |
Nanbeige4-3B-Thinking-2511 is an enhanced iteration over our previous Nanbeige4-3B-Thinking-2510.
|
| 25 |
Through advanced distillation techniques and reinforcement learning (RL) optimization, we have effectively scaled the model’s reasoning capacity, resulting in superior performance across a broad range of benchmarks.
|
| 26 |
On math and science reasoning benchmarks, Nanbeige4-3B-Thinking-2511 outperforms Qwen3-4B-Thinking-2507, Qwen3-8B-Thinking-2504, and Qwen3-14B-Thinking-2504 with a significant margin.
|
| 27 |
-
Besides, Nanbeige4-3B-Thinking-2511 achieves state-of-the-art (SOTA) results among models smaller than 32B parameters on Arena-Hard-V2 and BFCL-V4.
|
| 28 |
This marks a major milestone in delivering powerful, efficient reasoning performance at a compact scale.
|
| 29 |
|
|
|
|
|
|
|
| 30 |
<div align="center">
|
| 31 |
|
| 32 |
<img src="figures/performance_reasoning.png">
|
|
|
|
| 24 |
Nanbeige4-3B-Thinking-2511 is an enhanced iteration over our previous Nanbeige4-3B-Thinking-2510.
|
| 25 |
Through advanced distillation techniques and reinforcement learning (RL) optimization, we have effectively scaled the model’s reasoning capacity, resulting in superior performance across a broad range of benchmarks.
|
| 26 |
On math and science reasoning benchmarks, Nanbeige4-3B-Thinking-2511 outperforms Qwen3-4B-Thinking-2507, Qwen3-8B-Thinking-2504, and Qwen3-14B-Thinking-2504 with a significant margin.
|
| 27 |
+
Besides, Nanbeige4-3B-Thinking-2511 achieves state-of-the-art (SOTA) results among models smaller than 32B parameters on general tasks like Arena-Hard-V2 and BFCL-V4.
|
| 28 |
This marks a major milestone in delivering powerful, efficient reasoning performance at a compact scale.
|
| 29 |
|
| 30 |
+
* Technical Report - Comming Soon.
|
| 31 |
+
|
| 32 |
<div align="center">
|
| 33 |
|
| 34 |
<img src="figures/performance_reasoning.png">
|