UAE Lab Releases Open-Source Model to Rival China's DeepSeek (gizmodo.com)
- Reference: 0179210634
- News link: https://slashdot.org/story/25/09/13/1734225/uae-lab-releases-open-source-model-to-rival-chinas-deepseek
- Source link: https://gizmodo.com/uae-lab-releases-open-source-model-to-rival-chinas-deepseek-2000656197
"An Emirati AI lab called the Institute of Foundation Models (IFM) released K2 Think on Tuesday, a model that researchers say rivals OpenAI's ChatGPT and China's DeepSeek in standard benchmark tests."
> "With just 32 billion parameters, it outperforms flagship reasoning models that are 20x larger," the lab wrote in [1]a press release on Tuesday. DeepSeek's R1 has 671 billion parameters, though [2]only 37 billion are active . Meta's latest Llama 4 models range from [3]17 billion to 288 billion active parameters. OpenAI doesn't share parameter information. OpenAI doesn't share parameter information.
>
> Researchers also claim that K2 Think leads "all open-source models in math performance" across several benchmarks. The model is intended to be more focused on math, coding, and scientific research than most other AI chatbots. The Emirati lab's selling point for the model is similar to DeepSeek's strategy that [4]disrupted the AI market earlier this year: optimized efficiency that will have better or the same computing power at a lower cost...
>
> The lab is also aiming to be transparent in everything, "open-sourcing not just models but entire development processes" that provide "researchers with complete materials including training code, datasets, and model checkpoints," [5]IFM said in a press release from May .
The UAE and other Arab countries are investing in AI to try [6]reducing their economic dependence on fossil fuels , the article points out.
[1] https://mbzuai.ac.ae/news/mbzuai-and-g42-launch-k2-think-a-leading-open-source-system-for-advanced-ai-reasoning/
[2] https://www.ibm.com/think/topics/deepseek
[3] https://ai.meta.com/blog/llama-4-multimodal-intelligence/
[4] https://gizmodo.com/the-knives-are-coming-out-for-deepseek-ai-2000556375
[5] https://mbzuai.ac.ae/news/mbzuai-launches-institute-of-foundation-models-and-establishes-silicon-valley-ai-lab/
[6] https://www.bbc.com/news/articles/cvg8z2mx508o
PhallicGPT (Score:1)
PhallicGPT
Mantra? (Score:2)
OpenAI doesn't share parameter information. OpenAI doesn't share parameter information.
Re: (Score:2)
Or information on the murder of their whistleblower.
What is it with that "K2" naming? (Score:2)
I mean, we only recently heard about another LLM named "K2" from China: [1]https://developers.slashdot.or... [slashdot.org]
Is there something special about "K2" that I am not aware of?
[1] https://developers.slashdot.org/story/25/07/14/1942209/chinas-moonshot-launches-free-ai-model-kimi-k2-that-outperforms-gpt-4-in-key-benchmarks
Re: (Score:2)
It's "the second-highest mountain on Earth".
[1]https://en.wikipedia.org/wiki/... [wikipedia.org]
[1] https://en.wikipedia.org/wiki/K2
Re: (Score:2)
> It's "the second-highest mountain on Earth". [1]https://en.wikipedia.org/wiki/... [wikipedia.org]
Certainly, but why would you opt for such a name knowing others named their LLM also "K2"? They could have been more creative, and named it "Everest" or "LLM Khalifa", or whatever.
[1] https://en.wikipedia.org/wiki/K2
This is a good thing (Score:2)
Having one country or monopolist own the tech is bad
The best course of action is for the tech to be shared by all, preferably open source and open weight
Monopoly is inevitable (Score:2)
That's because the training data mostly comes from public internet sites.
And two things are going to destroy that training data.
First sights are gradually walking down to block AI bots if only because the excess traffic is crushing them.
And second the internet is filling up with AI slop and if you train your AI on slop you're going to get limitless slop
This means that before long the only people who will be able to maintain useful llms outside of a handful of extremely specific scientific purp
Flawed Reasoning (Score:3)
I notice they cite "math performance" as being a bright spot. Presumably they aren't talking about speed, but correctness.
Let that sink in: We now have to question not only how quickly our computers can perform a calculation, but whether the calculation is performed correctly at all .
I also notice the term "reasoning model" being applied. Everything I've read indicates the "reasoning" is just a differently-weighted LLM layered on top of the existing one. Artifice on top of artifice, with no real logical anchor. It sounds like exactly the right way to wave your hand for maximum voodoo, though.
Re: (Score:2)
> Let that sink in: We now have to question not only how quickly our computers can perform a calculation, but whether the calculation is performed correctly at all .
Yeah, just like with those "quantum computers". Maybe by running some LLM on a "quantum computer" we can have incorrect incorrect results... so, maybe correct ones?
Re: (Score:2)
Schroedinger's techbro. Until his answer is actually tested, it exists in a superposition of states. It is both right and wrong, charming and strange, confident and worthless.
Where's the link? (Score:2)
[1]https://chat.deepseek.com/ [deepseek.com] is for deepseek
Where's the link for K2?
I searched and it seems like it's only available for pay by other websites?
[1] https://chat.deepseek.com/
Re: (Score:1)
Everybody seems to hate some group these days.