Update README.md
Browse files
README.md
CHANGED
|
@@ -21,7 +21,7 @@ The Cogito v2 LLMs are instruction tuned generative models. All models are relea
|
|
| 21 |
- The LLMs are trained using **Iterated Distillation and Amplification (IDA)** - an scalable and efficient alignment strategy for superintelligence using iterative self-improvement.
|
| 22 |
- The models have been optimized for coding, STEM, instruction following and general helpfulness, and have significantly higher multilingual, coding and tool calling capabilities than size equivalent counterparts.
|
| 23 |
- In both standard and reasoning modes, Cogito v2-preview models outperform their size equivalent counterparts on common industry benchmarks.
|
| 24 |
-
-
|
| 25 |
|
| 26 |
# Evaluations
|
| 27 |
For detailed evaluations, please refer to the [Blog Post](https://www.deepcogito.com/research/cogito-v2-preview).
|
|
|
|
| 21 |
- The LLMs are trained using **Iterated Distillation and Amplification (IDA)** - an scalable and efficient alignment strategy for superintelligence using iterative self-improvement.
|
| 22 |
- The models have been optimized for coding, STEM, instruction following and general helpfulness, and have significantly higher multilingual, coding and tool calling capabilities than size equivalent counterparts.
|
| 23 |
- In both standard and reasoning modes, Cogito v2-preview models outperform their size equivalent counterparts on common industry benchmarks.
|
| 24 |
+
- This model is trained in over 30 languages and supports a context length of 128k.
|
| 25 |
|
| 26 |
# Evaluations
|
| 27 |
For detailed evaluations, please refer to the [Blog Post](https://www.deepcogito.com/research/cogito-v2-preview).
|