Join our daily and weekly newsletters for the latest updates and the exclusive content on AI coverage. Learn more
The story of the AI has reached a critical inflection point. THE Deep breakthrough – Achieving advanced performance without relying on the most advanced fleas – proves what many in Nerips in December have already declared: the future of AI does not consist in launching more calculation to problems – it is Acts to reinvent the way these systems work with humans with humans and our environment.
As a educated informator of Stanford who witnessed both the promise and the perils of AI development, I consider this moment as even more transformative than the beginnings of Chatgpt. We are entering what some call a “rebirth of reasoning”. O1 from OpenaiDeepseek’s R1, and others go beyond gross scaling towards something smarter – and doing it with unprecedented efficiency.
This change could not be more opportune. During his neirips arrangement, the former chief scientist of Openai Ilya Sutskever declared This “pre-training will end” because if the calculation power increases, we are limited by finished internet data. The breakthrough of Deepseek validates this perspective – the researchers of the Chinese company have obtained performances comparable to the O1 of Openai at a fraction of the cost, demonstrating that innovation, not only the power of raw calculation, is the way follow.
AI advanced without massive pre-training
Global models intensify to fill this gap. Recent World Labs $ 230 million Increase To build AI systems that understand reality as humans are parallel to the approach of Deepseek, where their R1 model presents “Aha!” Moments – Stop to reassess problems as humans do. These systems, inspired by human cognitive processes, promise to transform everything from environmental modeling into human-AI interaction.
We see early victories: Meta’s recent update to their Ray-Ban intelligent glasses Allows continuous contextual conversations with AI assistants without wake words, alongside real -time translation. It is not only an update of functionalities – it is an overview of how AI can improve human capacities without requiring massive pre -formulated models.
However, this evolution is accompanied by nuanced challenges. While Deepseek has considerably reduced costs thanks to innovative training techniques, this breakthrough of efficiency could paradoxically lead to an increase in overall consumption of resources – a phenomenon called Jevons paradoxWhere improvements in technological efficiency often lead to an increase rather than a reduction in resources.
In the case of AI, cheaper training could mean that more models are formed by more organizations, which could increase net energy consumption. But Deepseek’s innovation is different: by demonstrating that advanced performance is possible without advanced equipment, they do not only make AI more effective – they fundamentally change the way we approach the development of the model.
This change towards an intelligent architecture compared to the raw calculation power could help us to escape the trap of the Jevons paradox, while the accent is put on “how much calculation can we afford?” To “How can we intelligently design our systems?” As Professor Guy Den Broeck notes, Professor Guy Den Broeck, “the overall cost of the reasoning of the language model certainly does not drop.” The environmental impact of these systems remains substantial, pushing the industry to more effective solutions – exactly the type of Deepseek innovation represents.
Prioritize effective architectures
This change requires new approaches. Deepseek’s success validates the fact that the future is not to build more important models – it is a question of creating smarter and more effective models that work in harmony with human intelligence and environmental constraints.
Meta chief scientist Yann Lecun, considers future systems Spend days or weeks to think about complex problems, just like humans. The Deepseek’s-R1 model, with its ability to take a break and reconsider approaches, represents a step towards this vision. Although a resource delicacy, this approach could produce breakthroughs in climate change solutions, health care innovations and beyond. But like Carnegie Mellon Ameet Talwalkar Living wisely, we must question anyone claiming the certainty of the place where these technologies will lead us.
For business leaders, this change presents a clear path to follow. We must prioritize an effective architecture. The one who can:
- Deploy specialized AI agents channels rather than unique massive models.
- Invest in systems that optimize both for performance and environmental impact.
- Build infrastructure that supports iterative and human development in the loop.
Here is what excites me: the breakthrough of Deepseek proves that we pass beyond the era of “Big Bigger is Better” and in something much more interesting. With the pre-training of its limits and innovative companies to find new ways to achieve more with less, there is this incredible space open for creative solutions.
The intelligent chains of smaller and specialized agents are not only more effective – they help us to solve problems in a way that we would never have imagined. For startups and businesses ready to think differently, it is our moment to have fun again with AI, to build something that makes sense for people and the planet.
Kiara Nirghin is a technologist from Stanford Primé, successful author and co-founder of Chima.
DATADECISIONMAKERS
Welcome to the Venturebeat community!
Data data manufacturers are the place where experts, including technicians who do data work, can share data -related information and innovation.
If you want to read on advanced ideas and up-to-date information, best practices and the future of data and data technology, join us at datadecisionmakers.
You might even consider Contribute an article It’s up to you!