DeepSeek has released its new open-source model V3.1, with context length expanded to 128K

AI Daily News updated 8h ago dongdong
7 0

DeepSeek has announced the open-source release of its new base model, DeepSeek-V3.1-Base. After being published on Hugging Face, the model quickly rose to the 4th spot on the trending models list. DeepSeek-V3.1-Base adopts a Mixture of Experts (MoE) architecture, with its context length expanded to 128K, while maintaining the same number of parameters as the V3 version.

© Copyright Notice

Related Posts

No comments yet...

none
No comments yet...