India, Dec. 4 -- Chinese AI company with claDeepSeek proposed two new large language models, DeepSeek-V3.2 and DeepSeek-V3.2-Specialeimed performance roughly equal to or better than the recent proprietary frontier models (in both OpenAI and Google DeepMind), and which are open-source and more compute-efficient. The launch has excited the AI community in the world, particularly among people willing to develop powerful, yet affordable models.
DeepSeek characterises the novel models as developed based on the "Mixture-of-Experts" transformer architecture, which has approximately 671 billion total parameters. Nevertheless, active parameters are only a few, of the order 37 billion, per token when performing inference - a typical MoE approach t...
Click here to read full article from source
To read the full article or to get the complete feed from this publication, please
Contact Us.