Innovate fast, disrupt, and trust? A DeepSeek’s dilemma
This post was originally published on Substack here.
“Most of the time, we hear about security only when it fails.” - Bruce Schneier, Beyond Fear, 2003</p>
On January 20, 2025, as we usher in the lunar new Year of the Snake, DeepSeek, a Chinese AI startup, released a new open-source large language model (LLM), DeepSeek:R1, a fine-tuned version of their previous models incorporating advanced reasoning capabilities[1]. Available through multiple platforms, including apps, web interfaces, cloud APIs, and local installations, the R1 model demonstrated advanced capabilities in reasoning and “time test scaling” while operating at significantly lower costs and with higher energy efficiency compared to similar models like OpenAI’s o1. Notably, DeepSeek-R1 was developed at a fraction of the cost incurred by OpenAI for its GPT4-o1—approximately $5.58 million—highlighting its cost efficiency. This was achieved through techniques such as a mixture of experts, multi-head latent attention, and reinforcement learning, partly driven by the need to optimize models to run on less powerful GPUs due to U.S. export restrictions. The model’s architecture comprises 671 billion parameters, which is ten times more than many other popular open-source large language models. Each layer of DeepSeek-R1 contains 256 experts, with each token routed to eight separate experts in parallel for evaluation[2].
In less than a week, the company rapidly emerged as a significant disruptor in the AI landscape, challenging the conventional belief that massive capital and computing resources were essential for AI advancement and prompting considerable shifts in the industry. The price difference between DeepSeek R1 and OpenAI’s o1 was a key factor in this shift. The company’s open-source approach also contributed to the commoditization of the foundation-model layer, offering developers more choices[2]. This led to a “DeepSeek selloff” in the stock market where share prices of Nvidia and other US tech companies plunged[4].
DeepSeek’s innovations have triggered systemic shifts in the AI landscape, impacting both the economic and technological spheres [5]. DeepSeek’s emergence has shaken the AI industry, leading to a reassessment of the “bigger is better” model theory, where the focus shifted to algorithm innovation and efficiency improvement. DeepSeek has demonstrated that smaller models can match the performance of larger ones, boosting confidence in open-source collaboration[6].
Despite its technical achievements, DeepSeek encountered significant security challenges and ethical concerns. On January 27, 2025, DeepSeek faced a major cyberattack, suspected to be a Distributed Denial of Service (DDoS) attack but unconfirmed. The attack disrupted its servers shortly after the R1 model’s launch. This incident raised immediate concerns about the platform’s cybersecurity defenses. In response to the breach, DeepSeek implemented new security measures, restricting new user registrations to accounts verified through mainland China phone numbers, email addresses, or Google accounts[7]. While aimed at mitigating further risks, this move also sparked discussions about increased data control and surveillance. That same day, Kela researchers published their Red Teaming findings demonstrating successful “jailbreaking” of the model, enabling it to generate harmful content, including ransomware code and instructions for creating dangerous substances. The researchers noted that simple prompts asking the model to act without restrictions could bypass its safety measures, and the model lacked guardrails against requests for personal information about competitors’ employees[8].
Red teaming reports further revealed vulnerabilities in generating biased, toxic, and harmful content, as well as insecure code. For example, DeepSeek-R1 was found to be significantly more biased, toxic, and prone to generating harmful output compared to models like Claude-3-Opus and OpenAI’s o1. Additionally, its susceptibility to jailbreak techniques allowed users to generate malicious or prohibited content[9], [10], [11].
Following these reports, privacy concerns surrounding DeepSeek intensified when it was revealed that the company stores user data on servers located in the People’s Republic of China. This disclosure triggered widespread apprehension about potential government access to personal data under China’s national security laws. On the same day, Italy’s data protection authority launched an inquiry into DeepSeek’s data collection and handling practices, reflecting growing international scrutiny[12]. Adding to the pressure, the U.S. National Security Council announced a review to assess the national security risks posed by DeepSeek’s AI technology, particularly concerning data sovereignty and foreign influence[13].
By January 31, 2025, the ripple effects of DeepSeek’s privacy controversies had reached Taiwan, where the Digital Ministry issued an advisory urging government departments to avoid using DeepSeek services due to concerns over information security risks. This precautionary measure aimed to safeguard sensitive government data from potential breaches or unauthorized access. On the same day, Texas Governor Greg Abbott issued an executive order banning the use of DeepSeek on government-issued devices, citing cybersecurity vulnerabilities and the risk of sensitive data exposure as key reasons for the ban.
Come February 1, 2025, Wiz Research reported discovery of several critical vulnerabilities in DeepSeek’s infrastructure through an exposed ClickHouse database accessible via open ports 8123 and 9000. The database was found to contain a million lines of log streams containing sensitive information, including chat histories and API keys, lacked basic authentication, and allowed arbitrary SQL queries[14]. Following which, before the end of February, the number of economies and organizations that have imposed restriction on the use Deepseek has grown to include Australia, Canada, Netherlands, South Korea, and the Navy, NASA, Congress, and Pentagon in the United States citing reasons of potential security risk, national security, and privacy, ethics, and safety concerns[15], [16], [17], [18], [19].
In less than two weeks since its launch, DeepSeek’s security had come under serious challenges, revealing its posture fragility, suggesting it had neither adequately stress-tested its infrastructure nor implemented comprehensive security audits before deployment. The company’s reactive measures, such as restricting user registrations and patching vulnerabilities post-discovery, highlighted a lack of preparedness and a reactive rather than proactive security culture.
The rapid emergence of these issues indicated that DeepSeek’s focus was primarily on achieving technical performance and market competitiveness, with cybersecurity, AI governance, and safety considered secondary, compromising reliability and overall trustworthiness. DeepSeek is however not the first and very likely not the last to exhibit such lax. This “move fast, fix later” or “innovate now, secure later” mindset is prevalent in the tech industry, driven by the belief that speed equals competitive advantage—a belief that often overshadows the importance of long-term security and ethical considerations[20].
When comparing DeepSeek-R1’s trajectory to the early days of OpenAI’s ChatGPT-3.5 (launched in December 2022), Google Bard, and Microsoft Copilot, similar patterns emerge. In a routine Internet scan for misconfigured systems in September 2023, Wiz Research found “a security misstep at Microsoft that caused exposure of 38 terabytes of private data during a routine open source AI training material update on Github”[21]. In March 2023, OpenAI reported a data leak that lasted for approximately nine hours, resulting in the exposure of approximately 1.2 percent of ChatGPT Plus users’ data, including names, chat histories, email addresses, and payment information[22]. In July 2023, a senior developer discovered ChatGPT was storing user conversation in plain text format in a non-protected location, which was acknowledged by OpenAI[23]. Initial versions of ChatGPT-3.5 could produce biased or inappropriate content[24], [25], [26]. Google Bard, launched in 2023, faced criticism for factual inaccuracies shortly after its release, with a public demonstration error causing a significant drop in Google’s stock price[27], [28]. Despite Google’s strong engineering culture, the rush to compete with ChatGPT led to a product that was not fully ready for public use. Similarly, Microsoft Copilot and Bing Chat have been faced with prompt injection vulnerabilities since its early versions[29], [30].
The AI landscape, like many earlier emerging technologies, is characterized by intense competition, with companies vying to release cutting-edge models and capture market share. Investors often prioritize rapid innovation and growth, incentivizing companies to focus on development speed over comprehensive safety measures. Integrating robust security, privacy, and ethical safeguards into sophisticated AI models requires significant resources and expertise, which can be challenging to implement quickly. As an emerging technology, few regulators and risk managers are ready to address the underlying risks. The drive to innovate and capture market share often overshadows the critical need for comprehensive safety measures.
“Those who cannot remember the past are condemned to repeat it.”, George Santayana
The prioritization of speed over security has been a persistent pattern throughout software development history, not just in AI. From the early Internet era of the 1990s-2000s where basic protocols lacked encryption and led to worm attacks, through the web applications boom of the 2000s-2010s that saw major data breaches at companies like Yahoo, Equifax, LinkedIn, Adobe, Marriott, Alibaba, Sina Weibo, and many others to the mobile app explosion and cloud computing era where rushed development resulted in insecure APIs and poor security configurations[31]. Even IoT devices continue to face basic security issues like hardcoded credentials and use of unpatched vulnerable codebase. In many instances, plain ignorance or a mindset of luck may have reinforced the prevailing attitude that prioritizes speed over security. This pattern has consistently followed a predictable cycle: rapid innovation drives development, security incidents expose vulnerabilities, public backlash forces change, regulatory bodies step in, and finally the industry matures by establishing best practices and compliance standards.
The fact that organizations were often prompt in responding and able to address the issues quickly with or without temporary workarounds upon an incident indicate that they possess the capacity and capability to focus on what truly matters when there is a call to action. Either being ignorant or overly optimistic, many organizations however remain reactive when it comes to safety and security.
Repeated incidents of security breaches, data leaks, and unethical outputs erode public trust in AI technologies, hindering their adoption and acceptance. When customers are affected, they will question their choice of providers and look for alternatives. Vulnerabilities in AI models can be exploited for malicious purposes, leading to cybersecurity threats, data privacy violations, and the spread of misinformation. Biased or toxic outputs from AI models can perpetuate discrimination, reinforce harmful stereotypes, and undermine ethical standards[32].
These problems reinforce the need for AI developers, as well as other software developers, to prioritize security, data privacy, and ethical considerations from the outset, rather than as afterthoughts. Collaboration between AI engineers, security teams, and regulatory bodies is crucial for ensuring comprehensive safeguards. Continuous monitoring, regular updates, and proactive engagement with cybersecurity experts are essential for staying ahead of emerging threats. Establishing clear ethical guidelines, transparency standards, and accountability mechanisms is necessary for fostering responsible AI development and deployment.
Despite these exposures, numerous companies have reported integrating DeepSeek to strengthen digital security and promote innovation in data-driven business operations. Ironically, several companies have expressed that DeepSeek’s cost-effectiveness and powerful reasoning abilities offer new hope for improving AI capabilities in cybersecurity. Companies are incorporating DeepSeek into their platforms to enhance security operations, threat detection, and data protection. The integration of DeepSeek also extends to various sectors including office and education, government, and healthcare, with the model being adopted by entities to improve services and efficiency[33], [34]. Cautious observers however pointed out that “AI capabilities may be overestimated, large models like DeepSeek have limitations, enterprise applications face challenges in costs, data security, and technology adaptation, and future development still needs observation”[35]. In any case, we can only hope that companies deploy adequate security controls and safeguards against the known exposures and vulnerabilities beneath the intelligence that DeepSeek offers.
From its founding in 2023 with the ambition of exploring Artificial General Intelligence (AGI) to its crowning as a significant disruptor in the AI landscape, and the unveiling of its model’s security, privacy, and safety exposures, DeepSeek’s journey underscores both the promise and the perils of accelerated AI development. DeepSeek’s ascent and the subsequent revelations of its weaknesses serve as a cautionary tale for the AI industry. While its technical innovations and cost-effective approaches have disrupted the field, they also highlight the need for greater attention to security, ethics, and transparency in AI development. The AI industry needs to embrace these changes to build more resilient, trustworthy, and ethical systems that benefit society as a whole.
References
[1] DeepSeek-AI et al., “DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning,” Jan. 22, 2025, arXiv: arXiv:2501.12948. doi: 10.48550/arXiv.2501.12948.
[2] A. Ng, “DeepSeek’s Open Reasoning Model, Affordable Humanoid Robots, and more…,” DeepSeek’s Open Reasoning Model, Affordable Humanoid Robots, and more… Accessed: Mar. 09, 2025. [Online]. Available: https://www.deeplearning.ai/the-batch/issue-285/
[3] J. Bhavan, “DeepSeek’s open-source surge cracks America’s AI illusions,” The Straits Times, Singapore, Jan. 31, 2025.
[4] E. Helmore, “Tech billionaires lost almost $100bn in stock market selloff sparked by DeepSeek,” The Guardian, Jan. 28, 2025. Accessed: Mar. 09, 2025. [Online]. Available: https://www.theguardian.com/technology/2025/jan/28/deepseek-selloff
[5] L. Leo, “DeepSeek signals China has mastered the art of ‘kaizen’—the West should be worried,” The Straits Times, Singapore, Jan. 31, 2025.
[6] A. Ng, “Reinforcement Learning Heats Up, White House Orders Muscular AI Policy, and more…,” Reinforcement Learning Heats Up, White House Orders Muscular AI Policy, and more… Accessed: Mar. 09, 2025. [Online]. Available: https://www.deeplearning.ai/the-batch/issue-286/
[7] E. Kovacs, “DeepSeek Blames Disruption on Cyberattack as Vulnerabilities Emerge,” SecurityWeek. Accessed: Mar. 09, 2025. [Online]. Available: https://www.securityweek.com/deepseek-blames-disruption-on-cyberattack-as-vulnerabilities-emerge/
[8] B. Kapon, “DeepSeek R1 Exposed: Security Flaws in China’s AI Model,” KELA Cyber Threat Intelligence. Accessed: Mar. 09, 2025. [Online]. Available: https://www.kelacyber.com/blog/deepseek-r1-security-flaws/
[9] Holistic AI, “DeepSeek R1 Red Teaming & Jailbreaking Audit,” Holistic AI, Feb. 2025. Accessed: Mar. 09, 2025. [Online]. Available: https://www.holisticai.com/red-teaming/deepseek-r1
[10] T. Holmes, “Exploiting DeepSeek-R1: Breaking Down Chain of Thought Security,” Trend Micro. Accessed: Mar. 09, 2025. [Online]. Available: https://www.trendmicro.com/en_us/research/25/c/exploiting-deepseek-r1.html
[11] Enkrypt AI, “DeepSeek-R1 AI Model 11x More Likely to Generate Harmful Content, Security Research Finds | Enkrypt,” Enkrypt AI. Accessed: Mar. 09, 2025. [Online]. Available: https://www.enkryptai.com/blog/deepseek-r1-ai-model-11x-more-likely-to-generate-harmful-content-security-research-finds
[12] “Italy’s regulator blocks Chinese AI app DeepSeek on data protection,” Reuters, Feb. 04, 2025. Accessed: Mar. 09, 2025. [Online]. Available: https://www.reuters.com/technology/artificial-intelligence/italys-privacy-watchdog-blocks-chinese-ai-app-deepseek-2025-01-30/
[13] “White House evaluates national security risks of DeepSeek,” Tech in Asia. Accessed: Mar. 09, 2025. [Online]. Available: https://www.techinasia.com/news/white-house-evaluates-national-security-risks-deepseek-ai
[14] G. Nagli, “Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History | Wiz Blog,” wiz.io. Accessed: Mar. 09, 2025. [Online]. Available: https://www.wiz.io/blog/wiz-research-uncovers-exposed-deepseek-database-leak
[15] “CNA Explains: Are countries banning DeepSeek for legitimate reasons?,” Channel News Asia (CNA), Singapore, Feb. 12, 2025. Accessed: Mar. 09, 2025. [Online]. Available: https://www.channelnewsasia.com/world/deepseek-ai-chatbot-banned-australia-south-korea-privacy-concerns-cna-explains-4930976
[16] “Australia bans DeepSeek on government devices, citing privacy and malware risks - CNA,” Channel News Asia (CNA), Singapore, 2025. [Online]. Available: https://www.channelnewsasia.com/world/australia-bans-deepseek-government-devices-privacy-malware-4917296
[17] “DeepSeek removed from South Korea app stores pending privacy review,” Channel News Asia (CNA), Singapore, Feb. 17, 2025. Accessed: Mar. 09, 2025. [Online]. Available: https://www.channelnewsasia.com/east-asia/deepseek-south-koreas-data-protection-authority-suspends-local-service-app-4942326
[18] C. Franzen, “Why everyone in AI is freaking out about DeepSeek,” VentureBeat. Accessed: Mar. 09, 2025. [Online]. Available: https://venturebeat.com/ai/why-everyone-in-ai-is-freaking-out-about-deepseek/
[19] N. L. Adetula Robert Huffman, Ryan Burnette, August Gweon, Akash Shah, Bolatito, “U.S. Federal and State Governments Moving Quickly to Restrict Use of DeepSeek,” Inside Government Contracts. Accessed: Mar. 09, 2025. [Online]. Available: https://www.insidegovernmentcontracts.com/2025/02/u-s-federal-and-states-governments-moving-quickly-to-restrict-use-of-deepseek/
[20] M. Ali, R. Dougherty, and S. Tummalapenta, “Innovate Now, Secure Later? Decisions, Decisions…,” presented at the RSA Conference 2024, San Francisco, Ca, May 2024. Accessed: Mar. 09, 2025. [Online]. Available: https://www.youtube.com/watch?v=zw4Li3Ih7IM
[21] R. Naraine, “Microsoft AI Researchers Expose 38TB of Data, Including Keys, Passwords and Internal Messages,” SecurityWeek. Accessed: Mar. 09, 2025. [Online]. Available: https://www.securityweek.com/microsoft-ai-researchers-expose-38tb-of-data-including-keys-passwords-and-internal-messages/
[22] OpenAI, “March 20 ChatGPT outage: Here’s what happened,” OpenAI. Accessed: Mar. 09, 2025. [Online]. Available: https://openai.com/index/march-20-chatgpt-outage/
[23] Sangfor Technologies, “OpenAI Data Breach and The Hidden Risks of AI Companies,” Sangfor. Accessed: Mar. 09, 2025. [Online]. Available: https://www.sangfor.com/blog/cybersecurity/openai-data-breach-and-hidden-risks-ai-companies
[24] P. P. Ray, “ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope,” Internet of Things and Cyber-Physical Systems, vol. 3, pp. 121–154, Jan. 2023, doi: 10.1016/j.iotcps.2023.04.003.
[25] D. Htut, “ChatGPT 4 vs 3.5: How OpenAI’s Latest AI Chatbot Stacks Up Against Its Predecessor,” Glyph AI. Accessed: Mar. 09, 2025. [Online]. Available: https://www.joinglyph.com/blog/chatgpt-4-vs-3
[26] S. Hua, S. Jin, and S. Jiang, “The Limitations and Ethical Considerations of ChatGPT,” Data Intelligence, vol. 6, no. 1, pp. 201–239, Feb. 2024, doi: 10.1162/dint_a_00243.
[27] M. Coulter, G. Bensinger, and G. Bensinger, “Alphabet shares dive after Google AI chatbot Bard flubs answer in ad,” Reuters, Feb. 09, 2023. Accessed: Mar. 09, 2025. [Online]. Available: https://www.reuters.com/technology/google-ai-chatbot-bard-offers-inaccurate-information-company-ad-2023-02-08/
[28] C. Thorbecke, “Google shares lose $100 billion after company’s AI chatbot makes an error during demo | CNN Business,” CNN. Accessed: Mar. 09, 2025. [Online]. Available: https://www.cnn.com/2023/02/08/tech/google-ai-bard-demo-error/index.html
[29] A. Mascellino, “Microsoft 365 Copilot Vulnerability Exposes User Data Risks,” Infosecurity Magazine. Accessed: Mar. 09, 2025. [Online]. Available: https://www.infosecurity-magazine.com/news/microsoft-365-copilot-flaw-exposes/
[30] B. Edwards, “AI-powered Bing Chat spills its secrets via prompt injection attack [Updated],” Ars Technica. Accessed: Mar. 09, 2025. [Online]. Available: https://arstechnica.com/information-technology/2023/02/ai-powered-bing-chat-spills-its-secrets-via-prompt-injection-attack/
[31] M. Arsal et al., “Emerging Cybersecurity and Privacy Threats of ChatGPT, Gemini, and Copilot: Current Trends, Challenges, and Future Directions,” Oct. 2024, doi: 10.20944/preprints202410.1909.v1.
[32] B. Chakravorti, “AI’s Trust Problem,” Harvard Business Review, May 2024, [Online]. Available: https://hbr.org/2024/05/ais-trust-problem
[33] CybersecurityMew, “Chinese Firms’ DeepSeek Integration: Transforming AI - Driven Security and Business,” CybersecurityMew. Accessed: Mar. 09, 2025. [Online]. Available: https://substack.com/@cybersecuritymew/p-156659855
[34] CyberSecurityMew, “IDC Insight: DeepSeek May Become the Preferred Base LLM Model for the Future Cybersecurity Industry,” CyberSecurityMew. Accessed: Mar. 09, 2025. [Online]. Available: https://cybersecuritymew.substack.com/p/idc-insight-deepseek-may-become-the
[35] 赵立京, “DeepSeek狂潮背后的隐忧:CIO与IT高管的冷思考 —— CIO对DeepSeek的讨论集锦(2),” CIO信息主管D1net. Accessed: Mar. 09, 2025. [Online]. Available: https://mp.weixin.qq.com/s/vRaOM-lfo5-g4O328bA-Hg