More Powerful Than Meta and OpenAI: Chinese Startup DeepSeek Unveils AI Model

artificial intelligence

Chinese AI startup DeepSeek has introduced a new large language model that reportedly surpasses counterparts from Meta and OpenAI in testing.

The model, DeepSeek V3, boasts 671 billion parameters, compared to 405 billion in Llama 3.1. This indicates enhanced adaptability to complex applications and higher accuracy in responses.

image 151

The Hangzhou-based company trained the model in just two months with a budget of $5.58 million, using only 2,048 GPUs. This is significantly fewer resources than typically required by major tech firms. DeepSeek promises the best price-to-performance ratio in the market.

Future plans include introducing multimodality and “other advanced features.”

OpenAI team member Andrej Karpathy praised DeepSeek’s development, calling it impressive given the limited resources.

“This doesn’t mean large GPU clusters are unnecessary for cutting-edge LLMs, but it shows the importance of maximizing available resources. This project demonstrates there’s still much to optimize in both data and algorithms,” Karpathy added.

Previously, DeepSeek released a “competitor to OpenAI’s o1” — the advanced, “thinking” model DeepSeek-R1-Lite-Preview.

In July, Chinese company Kuaishou launched its video-generation AI model Kling, making it publicly available.