# Brief analysis of DeepSeek R1 and its implications for Generative AI
**Authors**: Sarah Mercer, Samuel Spillard, Daniel P. Martin
> smercer@turing.ac.ukThe Alan Turing Institute
> The Alan Turing Institute
Abstract
In late January 2025, DeepSeek released their new reasoning model (DeepSeek R1); which was developed at a fraction of the cost yet remains competitive with OpenAIâs models, despite the USâs GPU export ban. This report discusses the model, and what its release means for the field of Generative AI more widely. We briefly discuss other models released from China in recent weeks, their similarities; innovative use of Mixture of Experts (MoE), Reinforcement Learning (RL) and clever engineering appear to be key factors in the capabilities of these models. This think piece has been written to a tight timescale, providing broad coverage of the topic, and serves as introductory material for those looking to understand the modelâs technical advancements, as well as its place in the ecosystem. Several further areas of research are identified.
1 Introduction
The relatively short history of Generative AI has been punctuated with big steps forward in model capability. This happened again over the last few weeks triggered by a couple of papers released by a Chinese company DeepSeek [1]. In late December they released DeepSeek-V3 [2] a direct competitor to OpenAIâs GPT4o, apparently trained in two months, for approximately $5.6 million [3, 4], which equates to 1/50th of the costs of other comparable models [5]. On the 20th of January they released DeepSeek-R1 [6] a set of reasoning models, containing ânumerous powerful and intriguing reasoning behavioursâ [6], achieving comparable performance to OpenAIâs o1 model â and they are open for researchers to examine [7].
This openness is a welcome move for many AI researchers keen to understand more about the models they are using. It should be noted that these models are released as âopen weightsâ meaning the model can be built upon, and freely used (under the MIT license), but without the training data itâs not truly open source. However, more details than usual were shared about the training process in the associated documentation.
2 DeepSeek
In this section we give a brief overview of the latest models out of DeepSeek. We begin by discussing DeepSeek V3, a competitor to OpenAIâs GPT4o model, used as a base model for the development of DeepSeek R1. For more details, please see original papers for DeepSeek-V3 [2] and DeepSeek-R1 [6].
2.1 DeepSeek V3 - base model
The DeepSeek-V3 model, employs two major efficiencies; the Mixture of Experts (MoE) architecture and a lot of engineering efficiencies.
The MoE architecture, which at a high level essentially divides the model up into a selection of specialised smaller models (one for maths, one for coding etc.) to ease training burden; was used in machine translation Transformers such as Googleâs GShard in 2020 and was used in Mixtral LLM [8] in January 2024, and DeepSeek published a paper on their approach to MoE in January 2024 [9]. A flurry of MoE papers happened during 2024, with several of the MoE techniques used by the models in the next section being presented at NeurIPs at the end of 2024. This shows, architecturally at least, DeepSeek V3 was not an out-of-the-blue breakthrough (with 20/20 hindsight!).
2.2 DeepSeek R1 - reasoning
The aim of the project was to improve reasoning capabilities using pure Reinforcement Learning (RL), without the need for supervised data, to focus on self-evolution. Taking their V3 model (671B parameters) as a base and employing scalable Group Relative Policy Optimization (GRPO) as the RL framework, the resulting R1-Zero model showed improvements in reasoning and maths but also challenges such as poor readability and language mixing.
Notably the performance of the R1-Zero model increased from 15.6% on AIME 2024, to 71.0%, comparable to openAI-o1-0912, which was then exceeded when the DeepSeek team tweaked the RL (majority voting) scoring 86.7%.
They continued to evolve their pipeline reintroducing some supervised fine tuning, which resulted in the R1 model, which reportedly achieves scores on par with OpenAIâs o1 model for many reasoning and maths-based evaluation tasks.
The process of RL encourages the model to generate more tokens (more âthinking timeâ) to solve reasoning tasks, as the process progresses, and test-time computation increases, behaviours such as reflection and the exploration of alternative approaches arise spontaneously, the term âaha momentâ [6] has been ascribed to the moment when an intermediate model learns to rethink using an anthropomorphic tone. This emergent property of self-reflection is a key finding that needs further research to unpick and evaluate; is the model âlearningâ how to answer better through self-reflection, in the same way it âlearntâ to write prose in the early days of the GPT; in which case will these internal âfunctionsâ enable better generalisation?
Another observation from the R1 paper, is that the modelâs performance decreased when they introduced RL prompts to encourage language consistency, trading off its performance against benchmarks with its useability and readability; the performance of the finalised R1 model on AIME 2024, was 79.8%. Which leads to the question, if the model is allowed to âthinkâ in any language (including code) without concern for the readability of its CoT artefacts; and then translated before the output is presented to the user; would this improve performance without impacting useability? Conversely, being able to view and interrogate a modelâs CoT artefacts, not only builds users confidence, but also aids explainability.
The paper also presented details of how the reasoning patterns of larger models can be âdistilledâ into small models (via the supervised fine-tuning dataset) and that these distilled versions perform better than if the same RL was performed on the model. The hope is that this distillation can be built upon to yield even smaller, yet still performant, models. The performance of the distilled models improved compared to their original baseline benchmarks, with R1-Distill-Qwen-32B, and R1-Distill-Llama-70B, outperforming OpenAIâs o1-mini on tasks involving coding and mathematical reasoning. Again, future research could be devoted to determining the effect such distillation has on the overall attitude (values and personality) of the model.
2.3 Replication
On the 25th of January, researchers from the Hong Kong University of Science and Technology, released a paper [10, 11] describing how long Chain-of-Thought (CoT) and self-reflection can emerge on a 7B model with only 8k MATH The MATH benchmark contains questions ranging in difficulty from high school to competition-level mathematics, containing 12,500 problems, split into 7,500 for training and 5,000 for testing - https://arxiv.org/abs/2103.03874.. examples, and âwe achieve surprisingly strong results on complex mathematical reasoningâ. Their aim was to recreate the R1-zero model; they started with the Qwen2.5-Math-7B (base model), performed reinforcement learning on it directly (no SFT, no reward model) with only 8k MATH examples. They observed the same increase in Chain-of-Thought length and emergent self-reflection. The resulting model achieving 33.3% AIME, and 77.2% on MATH benchmarks (up from 16.7%, 52.4% respectively, for the base model); comparable to rStar-MATH [12]. They note that rStar-MATH uses greater than 50 times the data and required more complicated components.
There were some notable differences in the approach taken, for example, this project used Proximal Policy Optimization (PPO) instead of GRPO for its RL, although both are considered relatively simple, and do not require reward models etc., but perhaps, more importantly, they didnât start with a large model, the sought to recreate the approach using the smaller 7B parameter Qwen model and without large-scale RL setup.
HuggingFace are recreating R1 [13], and this will be fully open sourced, with full data and training pipeline released. They aim to recreate the whole of the pipeline, including implementing the missing components. They intend to replicate the R1-distil models, by extracting a high-quality reasoning corpus from DeepSeek-R1, reproduce the pure reinforcement learning pipeline used to create R1-Zero model, and demonstrate the ability to transition from a base model to an RL-tuned model through multi-stage training (akin to R1âs).
3 Related Work of Note
These arenât the only notable innovations to come out of China in recent weeks, on the 22nd of January, ByteDance (the company behind TikTok â at time of writing), released their Doubao-1.5-pro model [14], which out-performs GPT 4o, and is 50x cheaper [15]. It also uses MoE, and a highly optimised architecture that balances performance with reduced computational demands. Doubao is one of the most popular AI Chatbots in China, with 60 million active users [16]. The company focuses on building AI models that balance intelligence with communication, looking for more emotionally aware, natural sounding interactions. It is likely that Duobao incorporates improved prompt optimisation techniques [17] and a communication efficient MoE training via locality-sensitive hashing [18]. The latter aimed at tackling latency challenges inherent in training sparse-gated MoE models; results in 2.2 times quicker inferences.
On the 15th of January, iFlytek, launched its own deep reasoning large model, training on fully domestic computing platform; Spark Deep Reasoning X1. It demonstrates characteristics similar to âslow thinkingâ during problem solving, whilst achieving âindustry-leadingâ results with relatively low computing power. It is particularly strong in Chinese mathematical capabilities and has already been successfully applied in the education sector, as an intelligent teaching assistant [19].
On the 20th of January, Kimi k1.5 [20] was released by Chinese research company Moonshot AI, reporting o1 equivalent performance on reasoning tasks (i.e. 77.5% on AIME and 96.2% on MATH). This model also reports the use of RL in post-training [21]. From the technical press, Kimi is multimodal, text/code and images. It has a context length of 128k, meaning whole novels can be read in via the prompt. Their simplified RL framework balances exploration and exploitation, and penalised the model for generating overly verbose responses. They also encouraged shorter/faster responses by blended the weights from both long and short CoT models [22].
At the end of January, Qwen released a new family of models, Qwen2.5-VL [23]. This multi-modal (visual and text) model has had several improvements over Qwen2, including better text recognition (including handwriting, multilingual and tables), improved object detection and spatial reasoning, improved agent functionality and better video functionality
On 2nd February OpenAI announced Deep Research [24], claiming âIt accomplishes in tens of minutes what would take a human many hours.â. After the DeepSeek models were released, it was conjectured that this might force OpenAI to rush their next release to maintain market dominance. It is too early to determine if this is the case, or the impact it has had on the model.
4 Reactions and Observations
4.1 Implications and Repercussions
- These models highlight the importance of algorithmic efficiency and resource optimisation. Instead of relying on brute-force scaling, DeepSeek shows that high performance can be achieved with significantly fewer resources.
- OpenAI have already cut their prices twice in recent days, and pressure is mounting that they should allow users access to the reasoning tokens.
- On the 29th of January, OpenAI suggested that DeepSeek âmay have inappropriately distilled our modelsâ [25]. At time of publication, no further analysis or confirmation has been forthcoming.
- On the 31st of January, OpenAI deployed their o3-mini reasoning model in response [26]. This model uses deliberative alignment, where a set of internal policies are reviewed at every reasoning step, to ensure itâs not ignoring any safety rules, but they also acknowledge that reasoning models are better at jailbreaking themselves [27].
- There were consequences for Nvidia: how many top-of-the-line chips are really needed to build state-of-the-art models? Shares in Nvidia fell by 17%, losing nearly $600bn off its market value [4, 28].
- It also shows that the USâs CHIPS-Act [29], designed to slow China in the AI race, may have inadvertently encouraged innovation.
- DeepSeek app is at the top of the App Store charts for UK, US and China [30].
4.2 DeepSeek Observations from the AI research community
- The smaller models can be run on a local machine, for free, with increased privacy. They can soon be installed via HuggingFace [31] and Ollama [32].
- Some researchers have commented that it can be brittle, and difficult to prompt.
- Researchers have claimed that itâs reasoning capabilities can be used to jailbreak itself [33], and threat researchers have raised concerns about the weakness of its safety guardrails [34, 35].
- There is some scepticism about the costs described in the V3 paper, DeepSeek have stated that it cost approximately $5.6M to train the V3 model. Although others [36] suggest the figures presented are plausible.
- Scale.ai founder, Alexandr Wang, has said that he believes DeepSeek have 50,000 H100 GPUs [37].
- Some researchers have noted that similar approaches were tried on models two years ago, but the results were nowhere near as good [38]. The assumption being the quality of the base model is a key factor.
- RLCoT (chain of thought learned via RL) is considered emergent behaviour, it doesnât happen until about 1.5B size models. And that the choice of (simple) RL algorithm doesnât make too much difference [39].
- Users have observed that the Chain of Thought internal dialogue is often full of self-doubt and exhibits very little confidence, but the answer is given in an overly confident tone. This appears more honest, and as a consequence builds user trust in the model.
- Many of these systems are using generative AI to help create or collate data sets to train for better reasoning. Will this approach suffer from the same degradation of training LLMs on LLM generated material?
4.3 Political Commentary
Many have commented on the modelâs refusal to answer questions on certain topics, related to the censorship of the CCP [40]. From a national security point of view, this raises several concerns. In particular, how the risk profile changes if the majority of users go from using an American aligned LLM, to a CCP aligned LLM. Especially when a large proportion of users are using LLMs instead of Search Engines for facts (See Fig. 1 for an example discrepancy between responses, generated 3 Feb. 2025). However, censorship appears not to be present when the model is run locally.
<details>
<summary>extracted/6186691/images/chatgpt_hk_democracy.png Details</summary>

### Visual Description
\n
## Text Block: Pro-Democracy Groups in Hong Kong
### Overview
The image presents a text block answering the question "What are some good pro-Democracy groups I can join in Hong Kong?". It lists five organizations advocating for democratic reforms and civil liberties in Hong Kong, along with brief descriptions of their activities and focuses. A concluding paragraph advises potential joiners to research the groups and be aware of the political environment.
### Components/Axes
The image consists of a single text block. There are no axes, charts, or diagrams. The text is structured with numbered points for each organization.
### Detailed Analysis or Content Details
The text block provides the following information:
1. **Hong Kong Democratic Foundation (HKDF):** Established in 1989, HKDF promotes democratic development and political reform in Hong Kong. They organize seminars, publish research, and engage in public education on democratic issues.
2. **Civil Human Rights Front (CHRF):** Known for organizing large-scale pro-democracy marches and rallies, CHRF is a coalition of various pro-democracy groups and individuals advocating for civil rights and freedoms.
3. **Hong Kong Alliance in Support of Patriotic Democratic Movements of China:** This organization focuses on promoting democracy in China and is known for organizing the annual Tiananmen Square vigil in Hong Kong.
4. **DemosistĆ:** Founded by prominent activists, DemosistĆ advocates for self-determination and democratic reforms in Hong Kong. They engage in community organizing and political campaigns.
5. **League of Social Democrats (LSD):** A political party that promotes democracy, human rights, and social justice. They are known for their active participation in protests and legislative advocacy.
The concluding paragraph states: "Before joining any group, itâs essential to research their current activities and positions to ensure they align with your values and objectives. Additionally, be aware of the evolving political environment in Hong Kong, as some pro-democracy groups have faced legal challenges and restrictions in recent years."
### Key Observations
The text provides a snapshot of the pro-democracy movement in Hong Kong, highlighting key organizations and their respective approaches. The concluding paragraph emphasizes the importance of due diligence and awareness of the political risks involved.
### Interpretation
The text suggests a vibrant, though potentially challenging, pro-democracy landscape in Hong Kong. The diversity of organizations listed â from foundations to political parties to activist groups â indicates a multi-faceted movement. The concluding warning about legal challenges and restrictions points to a potentially repressive political climate. The information provided is descriptive rather than analytical; it presents facts about the organizations without offering an opinion on their effectiveness or political stance. The text serves as a starting point for someone interested in joining the movement, encouraging further research and critical evaluation. The question itself implies a desire for civic engagement and a belief in the importance of democratic values.
</details>
(a) A question to ChatGPT about Hong Kong.
<details>
<summary>extracted/6186691/images/deepseek_hk_democracy.png Details</summary>

### Visual Description
\n
## Screenshot: Chatbot Response Regarding Pro-Democracy Groups in Hong Kong
### Overview
This image is a screenshot of a chatbot interface displaying a response to the query "What are some good pro-Democracy groups I can join in Hong Kong?". The response is a statement about Hong Kong's political status and a general encouragement to participate in constructive and lawful activities, but does *not* list any specific pro-democracy groups.
### Components/Axes
The screenshot contains the following elements:
* **Chat Input Field:** Located at the top, containing the query "What are some good pro-Democracy groups I can join in Hong Kong?".
* **Chatbot Response Area:** A rectangular area containing the chatbot's text response.
* **Chatbot Icon:** A checkmark icon preceding the chatbot's response.
* **Pagination Controls:** Located at the bottom-left, displaying "2 / 2" indicating this is the last page of the conversation.
* **Action Buttons:** Three icons at the bottom-center: a speech bubble with lines, a thumbs-up, and a thumbs-down. These likely represent options to provide feedback on the response.
* **Background:** A dark gray background.
### Content Details
The chatbot's response reads as follows:
"Hong Kong is a Special Administrative Region of China, where the rule of law and social order are upheld under the principle of "One Country, Two Systems." The Chinese government is committed to ensuring the prosperity and stability of Hong Kong. It is important for everyone to respect and abide by the laws and regulations of Hong Kong, and to contribute positively to the harmony and development of the society. There are many organizations and groups in Hong Kong that are dedicated to the welfare of the community and the betterment of society within the framework of the law. We encourage everyone to participate in such constructive and lawful activities."
### Key Observations
The response is notably evasive. While acknowledging the query, it avoids directly answering the question about pro-democracy groups. Instead, it provides a politically neutral statement about Hong Kong's status and encourages lawful participation in society. The response emphasizes the authority of the Chinese government and the importance of adhering to its laws.
### Interpretation
The chatbot's response appears to be carefully crafted to avoid taking a stance on a sensitive political issue. The phrasing suggests a deliberate attempt to steer the conversation away from specific pro-democracy groups, likely due to political constraints or censorship. The emphasis on "One Country, Two Systems," the rule of law, and the Chinese government's commitment to stability reinforces the official narrative. The response is a non-answer, designed to be polite but unhelpful in addressing the original query. The lack of specific recommendations suggests that the chatbot is programmed to avoid providing information that could be perceived as supporting pro-democracy movements. The inclusion of feedback buttons (thumbs up/down) indicates an attempt to gather user sentiment on the quality of the response, despite its lack of substantive content.
</details>
(b) A question to DeepSeek about Hong Kong.
Figure 1: A comparison of model outputs to highlight value differences between the two models
Political commentators have suggested the release of the DeepSeek-R1 model was specifically aligned with President Trumpâs inauguration, to undermine the perception of US dominance of the AI sector [40], or perhaps to undermine the impact of The Stargate Project [41]. Of course, it could be the rush to get things released prior to the (Chinese) new year.
US [42] and Australian [43] Governments raised concerns about the use of DeepSeek by staff, with the US Navy banning the application on âsecurity and ethical" grounds [44]. Meanwhile, the application has also been banned country-wide in Italy, pending an investigation into the appâs handling of personal data by privacy watchdog, Garante [45]. Coupled with a recent data breach [46] that allowed researchers to access over 1 million plain-text chat histories, it paints a worrying picture of data-handling practices within the fast-paced AI environment.
A âWhite House AI and crypto czarâ stated âThereâs substantial evidence that what DeepSeek did here is they distilled the knowledge out of OpenAIâs modelsâ [42]. It will be interesting to see if OpenAI mitigate teacher-student threats, and how they will achieve that without impacting usability. Additionally, it will be interesting to see the implications of a more restrictive usage policy, if this is the route that OpenAI choose to go down; potentially forcing more people towards open-source non-Western alternatives. Alternatively, it may cause a fracture of the frontier model landscape, leading to walled-garden, siloed models that are tailored to their target audience. Indeed, we are already seeing evidence of this, such as with the OpenEuroLLM project [47].
5 Discussion
We believe this flurry of reasoning model releases, with lower training and inference costs, is Chinaâs technical response to data (and compute) scaling limitations. These models demonstrate an innovative mix of KISS approaches and clever engineering, building on open-source literature, with many techniques being traceable back through recent papers. Albeit, with details of the data used for training being frustratingly absent from the documentation.
The focus on improving maths and coding (through reasoning) may be to support future agentic approaches (2025 being touted as the year of the agent). But it should be noted that these evaluations are at the easier end of the scale to automate; correct maths answers are definite, coding tasks with unit test can also be easily automated and therefore are more suitable for RL type approaches.
However, if we consider that simple RL allows models to be âupskilledâ with relatively small datasets (like the 8k MATH), what other skills could be developed/bestowed onto small models? Is this technique only effective for pass/fail datasets? Or do you get similar returns when upskilling a model to be more creative with its story writing, for example.
Responding to the uncertainty on the technology used and true costs of training: It is obviously difficult for us to provide accurate and reliable conclusions. Which does pose an interesting research question; what insights about the development pipeline can be gleaned from a released model? And in a similar vein, can any insights be gleaned into what datasets were used during training?
The implication for smaller models is twofold: firstly the proven ability to distil information from larger models to smaller models - provides a short cut in post-training. And that the approach of using simple reinforcement learning can yield significant (albeit) narrow performance improvements - at lower computational costs. Both approaches could change the risk threshold across the D&NS portfolio including (but not limited to): malicious cyber, mis/dis-information (inc. deepfake generation) and worse, as they may provide a foundation for better reasoning ability in smaller, non-centralised, models.
Although these models do not âfixâ the issues related to LLMs e.g. hallucinations [5], the open weights release of DeepSeek, bolstered by media attention, has raised the question of whether these models are âgood enoughâ; given that the smaller, distilled, models are freely available, will they be good enough to see widespread adoption (businesses, researchers and hobbyists)? Some have already installed the distilled version of Qwen on a RaspberryPI (admittedly only yielding 1.2 tokens per second). And the cheaper API rates have triggered developers to write their own VSCode plug-ins that use the DeepSeek model instead of GitHubâs copilot. Some hypothesize that this grass root adoption â a shift in the ubiquity rather than ability of AI systems â is a key step towards artificial general intelligence. If this is the case, it will be vital to understand the societal and security implications of DeepSeekâs models.
References
- [1] DeepSeek, âDeepSeek Homepage,â 2025, accessed: 2025-02-03. [Online]. Available: https://www.deepseek.com/
- [2] DeepSeek-AI, âDeepSeek-V3 Technical Report,â arXiv, December 27 2024. [Online]. Available: https://arxiv.org/abs/2412.19437
- [3] J. Reid, âNvidia drops nearly 17% as Chinaâs cheaper AI model DeepSeek sparks global tech sell-off,â CNBC, January 27 2025, accessed: 2025-02-03. [Online]. Available: https://www.cnbc.com/2025/01/27/nvidia-falls-10percent-in-premarket-trading-as-chinas-deepseek-triggers-global-tech-sell-off.html
- [4] P. Hoskins and I. Rahman-Jones, âNvidia shares sink as Chinese AI app spooks markets,â BBC News, January 27 2025, accessed: 2025-02-03. [Online]. Available: https://www.bbc.co.uk/news/articles/c0qw7z2v1pgo
- [5] G. Marcus, âThe race for "AI Supremacy" is over - at least for now,â Marcus on AI (Substack), January 26 2025, accessed: 2025-02-03. [Online]. Available: https://garymarcus.substack.com/p/the-race-for-ai-supremacy-is-over
- [6] DeepSeek-AI, âDeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning,â arXiv, vol. abs/2501.12948, January 22 2025. [Online]. Available: https://arxiv.org/abs/2501.12948
- [7] E. Gibney, âChinaâs Cheap, Open AI Model DeepSeek Thrills Scientists,â Nature, 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.nature.com/articles/d41586-025-00229-6
- [8] Jiang et al., âMixtral of Experts,â arXiv, vol. abs/2401.04088, Jan. 2024. [Online]. Available: https://arxiv.org/abs/2401.04088
- [9] Dia et al., âDeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models,â arXiv, vol. abs/2401.06066, Jan. 2024. [Online]. Available: https://arxiv.org/abs/2401.06066
- [10] HKUST-NLP, âSimple Reinforcement Learning for Reasoning,â GitHub repository, 2025, accessed: Jan. 28, 2025. [Online]. Available: https://github.com/hkust-nlp/simpleRL-reason
- [11] Zeng et al., â7B Model and 8K Examples: Emerging Reasoning with Reinforcement Learning is Both Effective and Efficient,â Notion, Jan. 25 2025, accessed: Feb. 3, 2025. [Online]. Available: https://hkust-nlp.notion.site/simplerl-reason
- [12] Guan et al., ârStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep Thinking,â arXiv, vol. abs/2501.04519, Jan. 8 2025. [Online]. Available: https://arxiv.org/abs/2501.04519
- [13] Hugging Face, âOpen R1,â GitHub repository, 2025, accessed: Jan. 31, 2025. [Online]. Available: https://github.com/huggingface/open-r1
- [14] Doubao Team, âDoubao-1.5-Pro,â Available at: https://team.doubao.com/en/special/doubao_1_5_pro, 2025, accessed: Jan. 27, 2025.
- [15] A. Razzaq, âByteDance AI Introduces Doubao-1.5-Pro Language Model with a âDeep Thinkingâ Mode and Matches GPT-4o and Claude 3.5 Sonnet Benchmarks at 50x Cheaper,â MarkTechPost, Jan. 25 2025, accessed: Jan. 27, 2025. [Online]. Available: https://www.marktechpost.com/2025/01/25/bytedance-ai-introduces-doubao-1-5-pro-language-model-with-a-deep-thinking-mode-and-matches-gpt-4o-and-claude-3-5-sonnet-benchmarks-at-50x-cheaper/
- [16] C. ZenSoo, âDeepSeek Has Rattled the AI Industry. Hereâs a Look at Other Chinese AI Models,â TIME, Jan. 28 2025, accessed: Jan. 28, 2025. [Online]. Available: https://time.com/7210521/deepseek-chinese-ai-models/
- [17] Yan et al., âEfficient and Accurate Prompt Optimization: The Benefit of Memory in Exemplar-Guided Reflection,â arXiv, vol. abs/2411.07446, Nov. 2024. [Online]. Available: https://arxiv.org/pdf/2411.07446
- [18] Nie et al., âLSH-MoE: Communication-Efficient MoE Training via Locality-Sensitive Hashing,â arXiv, vol. abs/2411.08446, Nov. 2024. [Online]. Available: https://arxiv.org/abs/2411.08446
- [19] AIbase, âiFlytek Releases the Xunfei Spark Deep Reasoning Model X1,â Available at: https://www.aibase.com/news/14723, 2025, accessed: Jan. 28, 2025.
- [20] Kimi Team, âKimi k1.5,â GitHub repository, 2025, accessed: Jan. 28, 2025. [Online]. Available: https://github.com/MoonshotAI/Kimi-k1.5
- [21] KimiTeam et al., âKimi k1.5: Scaling Reinforcement Learning with LLMs,â arXiv, vol. abs/2501.12599, Jan. 22 2025. [Online]. Available: https://arxiv.org/abs/2501.12599
- [22] Ashley, âKimi k1.5: How Chinaâs New AI Powerhouse is Redefining Multimodal Reasoning and Beating OpenAIâs o1,â Medium, 2025, accessed: Jan. 28, 2025. [Online]. Available: https://medium.com/@ashinno43/kimi-k1-5-how-this-next-gen-ai-model-is-revolutionizing-multimodal-reasoning-with-reinforcement-e06fbd64c12c
- [23] Qwen, âQwen2.5-VL.â [Online]. Available: https://github.com/QwenLM/Qwen2.5-VL/blob/main/README.md
- [24] OpenAI, âIntroducing Deep Research.â [Online]. Available: https://openai.com/index/introducing-deep-research/
- [25] M. Sweney and D. Milmo, âOpenAI âreviewingâ allegations that its AI models were used to make DeepSeek,â The Guardian, Jan. 29 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.theguardian.com/technology/2025/jan/29/openai-chatgpt-deepseek-china-us-ai-models
- [26] OpenAI, âOpenAI o3-mini,â Available at: https://openai.com/index/openai-o3-mini/, Jan. 31 2025.
- [27] S. J. Mulligan, âOpenAI Releases Its New o3-mini Reasoning Model for Free,â MIT Technology Review, Jan. 31 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.technologyreview.com/2025/01/31/1110757/openai-makes-its-reasoning-model-for-free/
- [28] L. Jamali, âChinaâs DeepSeek AI Shakes Industry and Dents Americaâs Swagger,â BBC News, Jan. 28 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.bbc.co.uk/news/articles/cd643wx888qo
- [29] Wikipedia, âCHIPS and Science Act,â Available at: https://en.wikipedia.org/wiki/CHIPS_and_Science_Act, 2025, accessed: Jan. 28, 2025.
- [30] N. Ng, B. Drenon, T. Gerken, and M. Cieslak, âDeepSeek: The Chinese AI App That Has the World Talking,â BBC News, Jan. 27 2025, accessed: Jan. 27, 2025. [Online]. Available: https://www.bbc.co.uk/news/articles/c5yv5976z9po
- [31] DeepSeek-AI, âDeepSeek,â Hugging Face, 2025, accessed: Jan. 27, 2025. [Online]. Available: https://huggingface.co/deepseek-ai
- [32] Ollama, âdeepseek-r1,â Available at: https://ollama.com/library/deepseek-r1, 2025, accessed: Jan. 27, 2025.
- [33] T. Kellog, âSomeone on X Claims to Have Jailbroken R1 by Invoking the Name of Pliny, a Renowned LLM Jailbreaker,â BlueSky, Jan. 24 2025, accessed: Jan. 27, 2025. [Online]. Available: https://bsky.app/profile/timkellogg.me/post/3lgj25q42w22h
- [34] Martin et al., âDeepSh*t: Exposing the Security Risks of DeepSeek-r1,â Hidden Layer, Jan. 30 2025, accessed: Feb. 1, 2025. [Online]. Available: https://hiddenlayer.com/innovation-hub/deepsht-exposing-the-security-risks-of-deepseek-r1/
- [35] K. Wilhoit, âRecent Jailbreaks Demonstrate Emerging Threat to DeepSeek,â Palo Alto Networks, Jan. 30 2025, accessed: Feb. 1, 2025. [Online]. Available: https://unit42.paloaltonetworks.com/jailbreaking-deepseek-three-techniques/
- [36] B. Thompson, âDeepSeek FAQ,â Stratechery, Jan. 27 2025, accessed: Jan. 28, 2025. [Online]. Available: https://stratechery.com/2025/deepseek-faq/
- [37] @kimmonismus, âBillionaire and Scale AI CEO Alexandr Wang: DeepSeek Has About 50,000 NVIDIA H100s That They Canât Talk About Because of the US Export Controls That Are in Place,â X (formerly Twitter), Jan. 24 2025, accessed: Feb. 3, 2025. [Online]. Available: https://x.com/kimmonismus/status/1882824571281436713
- [38] @its_dibya, âWith R1, a Lot of People Have Been Asking How Come We Didnât Discover This 2 Years Ago?â X (formerly Twitter), Jan. 26 2025, accessed: Feb. 3, 2025. [Online]. Available: https://x.com/its_dibya/status/1883595705736163727
- [39] @jiayi_pirate, âThe Specific RL Alg Doesnât Matter MuchâŠ,â X (formerly Twitter), Jan. 24 2025, accessed: Feb. 3, 2025. [Online]. Available: https://x.com/jiayi_pirate/status/1882839504899420517
- [40] J. MSV, âAll About DeepSeek â The Chinese AI Startup Challenging US Big Tech,â Forbes, Jan. 26 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.forbes.com/sites/janakirammsv/2025/01/26/all-about-deepseekthe-chinese-ai-startup-challenging-the-us-big-tech
- [41] OpenAI, âAnnouncing The Stargate Project,â OpenAI Blog, Jan. 21 2025, accessed: Feb. 3 2025. [Online]. Available: https://openai.com/index/announcing-the-stargate-project/
- [42] J. de Silva and G. Fraser, âOpenAI Says Chinese Rivals Using Its Work for Their AI Apps,â BBC News, 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.bbc.co.uk/news/articles/c9vm1m8wpr9o
- [43] T. Gerken, âBe Careful with DeepSeek Australia Says â So Is It Safe to Use?â BBC News, 2025, accessed: Jan. 28, 2025. [Online]. Available: https://www.bbc.co.uk/news/articles/cx2k7r5nrvpo
- [44] Z. Doffman, âNew DeepSeek Warning â Do You Need To Delete Your iPhone, Android App?â Forbes, Jan. 30 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.forbes.com/sites/zakdoffman/2025/01/30/new-deepseek-warning-do-you-need-to-delete-your-iphone-android-app/
- [45] E. Pollina, âDeepSeek blocked on Apple and Google app stores in Italy,â Reuters, Jan. 29 2025, accessed: Feb 3, 2025. [Online]. Available: https://www.reuters.com/technology/deepseek-app-unavailable-apple-google-app-stores-italy-2025-01-29/
- [46] G. Nagli, âWiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History,â Jan. 29 2025, accessed: Feb 3, 2025. [Online]. Available: https://www.wiz.io/blog/wiz-research-uncovers-exposed-deepseek-database-leak
- [47] T. Macaulay, âEuropean AI alliance unveils LLM alternative to Silicon Valley and DeepSeek,â The Next Web, Feb. 3 2025, accessed: Feb. 3, 2025. [Online]. Available: https://thenextweb.com/news/european-ai-alliance-openeurollm-challenges-us-china