February, is rumored to outperform ChatGPT and Claude in long-context coding, targeting elite-level coding tasks.
DeepSeek founder Liang Wenfeng has published a new paper with a research team from Peking University, outlining key technical ...
DeepSeek has released a new AI training method that analysts say is a "breakthrough" for scaling large language models.
DeepSeek's new research enables retrieval using computational memory, not neural computation, freeing up GPUs.
Another Chinese quantitative trading firm has entered the race to develop large language models (LLMs), unveiling systems it ...
DeepSeek has released new research showing that a promising but fragile neural network design can be stabilised at scale, ...
Chinese and Western large language models are reshaping global information power, embedding political world views into the ...
The Chinese AI lab may have just found a way to train advanced LLMs in a manner that's practical and scalable, even for more cash-strapped developers.
DeepSeek's latest technical paper, co-authored by the firm's founder and CEO Liang Wenfeng, has been cited as a potential ...
Chinese AI startup DeepSeek is expected to launch its next-generation AI model that features strong coding capabilities in ...
DeepSeek's proposed "mHC" architecture could transform the training of large language models (LLMs) - the technology behind ...
DeepSeek continues to push the frontier of generative AI...in this case, in terms of affordability. The company has unveiled its latest experimental large language model (LLM), DeepSeek-V3.2-Exp, that ...