Description
DeepSeek is an open LLM model promising innovation in both efficiency and transparency – but how much do we really know about what’s happening under the hood? And what does it tell us about where we are in AI development?
In this session, we examine DeepSeek from two perspectives:
Security and risk: What challenges arise when an AI model is built and distributed openly? What does this mean for reliability, integrity, and control? What risks come with DeepSeek’s Chinese origins, and how does that affect trust in the model?
Technical architecture and efficiency: Has DeepSeek truly broken new ground, or is it just another example of solving inefficiency by throwing more hardware at the problem?
We’ll explore the technical details of Mixture of Experts (MoE), parallelization strategies, and why LLMs still resemble steam engines – powerful but wasteful. We’ll also discuss a fundamental limitation of today’s transformer-based AI: that a model’s performance is closely tied to data availability, which in turn dictates how large and effective it can actually become. Finally, we’ll look at why the AI industry – DeepSeek included – remains so reluctant to be transparent about training methods and optimization.
Join us for a deep dive into both the risks and possibilities of open-source LLMs!
Optional: Speaker / convener biography
Omegapoint, VP Academy, General AI, Head of Innovation and Advocates
Programmer with an interest for security. Uses high-quality and low-latency development to drive security. Agile aficionado, DDD enthusiast, and DevOps admirer. Dabbled with AI since 1997. Author of "Secure by Design" [Manning], contributing author to "97 Things every Programmer should Know" and "LESS! Essays on Business Transformation". Speaker.
Length | 45 minutes |
---|