Berkeley Researchers Recreate DeepSeek AI Technology for $30

Extended summary

Published: 31.01.2025

Introduction

Researchers at the University of California, Berkeley, have successfully recreated the core technology of China's DeepSeek AI for a remarkably low cost of just $30. This achievement signals a potential shift in the AI landscape, suggesting that advanced models can be developed at a fraction of the expense typically associated with leading technology firms. Led by Ph.D. candidate Jiayi Pan, the study highlights the capabilities of smaller models in reinforcement learning, challenging the notion that only large-scale investments yield significant advancements in artificial intelligence.

Recreation of DeepSeek Technology

The Berkeley team focused on replicating the reinforcement learning features of DeepSeek R1-Zero using a small language model with 3 billion parameters. Despite its smaller size compared to industry counterparts, the model exhibited impressive self-verification and search capabilities, which are essential for refining responses. The researchers tested their recreation through a numerical puzzle known as the Countdown game, where the AI initially struggled with random guesses. However, through reinforcement learning, it progressively developed self-correction techniques and improved its problem-solving abilities, ultimately arriving at the correct answers.

Model Testing and Performance

To evaluate the effectiveness of their AI model, the researchers also examined its multiplication skills. The AI demonstrated an ability to decompose equations using the distributive property, mirroring the cognitive strategies humans use to tackle complex multiplication problems. The research team experimented with various model sizes, starting with a 500-million-parameter version that lacked accuracy. As they scaled the model up to 1.5 billion parameters, they noticed the incorporation of revision techniques, while models ranging from 3 to 7 billion parameters exhibited significant improvements in problem-solving efficiency and accuracy.

Cost Comparison and Implications

The cost effectiveness of the Berkeley team's approach becomes even more striking when compared to leading AI companies. For context, OpenAI charges approximately $15 per million tokens for its API, while DeepSeek offers a lower rate of about $0.55 per million tokens. The findings suggest that advanced AI models can be developed with far less financial investment than what is typically seen in the industry. This raises important questions about the sustainability and scalability of current AI development practices, particularly in light of the substantial budgets allocated by major firms like OpenAI, Google, and Microsoft.

Concerns and Skepticism

Despite the promising results from the Berkeley team, there are notable concerns regarding the DeepSeek AI's reliability and ethical implications. Some experts, such as AI researcher Nathan Lambert, have expressed skepticism about DeepSeek's claimed affordability. Lambert questions the accuracy of the reported $5 million training cost for its 671-billion-parameter model and suggests that operational expenses could be significantly higher when considering infrastructure and research costs. Additionally, the AI's data transmission back to China has raised privacy concerns, leading to bans in various regions, including the U.S.

Conclusion

The work conducted by the Berkeley researchers underscores the potential for developing advanced AI technologies with minimal financial investment. Their findings challenge the prevailing notion that only large-scale operations can produce cutting-edge models. As the AI field continues to evolve, this research may pave the way for more accessible and cost-effective solutions, while also prompting critical discussions about the ethical and operational standards within the industry.

Source: BGR

Top Headlines 31.01.2025