Chinese AI startup DeepSeek has unveiled DeepSeek Coder V2, an open-source mixture of experts (MoE) code language model that promises to outperform state-of-the-art closed models like GPT-4 Turbo, Claude 3 Opus, and Gemini 1.5 Pro. Built on the foundation of the DeepSeek-V2 model and trained on a massive dataset of 6 trillion tokens, DeepSeek Coder V2 supports over 300 programming languages, making significant strides in coding and math tasks. It features a context window of 128K, allowing it to handle more complex coding challenges. The model has shown superior performance across various benchmarks, establishing itself as a leader in the open-source domain. DeepSeek’s approach uses a Mixture of Experts framework to activate specific parameters for optimized performance. Additionally, the model maintains strong general language and reasoning capabilities. DeepSeek Coder V2 is available under an MIT license, promoting both research and commercial use.

Source.

TOP STORIES

Unauthorized Users Breach Anthropic's Mythos Cybersecurity Tool
Unauthorized users have gained access to Anthropic’s Mythos, raising security concerns …
Clarifai Deletes 3 Million Photos Amid FTC Investigation Over Data Use
Clarifai has deleted millions of photos from OkCupid amid an FTC investigation into data misuse …
Nvidia's AI Revolution - The Vera Rubin Platform and Future Demand
Nvidia’s Vera Rubin platform is set to revolutionize AI inference with unmatched performance …
Tim Cook's Departure Marks a New Era for Apple's AI Strategy
Apple’s leadership changes signal a strategic shift towards AI and silicon innovation …
Tim Cook's Departure - A Strategic Shift in Apple's AI Landscape
Apple’s leadership transition highlights a strategic focus on silicon for AI innovation …
New Tennessee Law on AI and Mental Health - A Step Forward or Backward?
Tennessee’s new law restricts AI claims in mental health but may create loopholes …

latest stories