# Brief analysis of DeepSeek R1 and its implications for Generative AI
**Authors**: Sarah Mercer, Samuel Spillard, Daniel P. Martin
> The Alan Turing Institute
## Abstract
In late January 2025, DeepSeek released their new reasoning model (DeepSeek R1); which was developed at a fraction of the cost yet remains competitive with OpenAIâs models, despite the USâs GPU export ban. This report discusses the model, and what its release means for the field of Generative AI more widely. We briefly discuss other models released from China in recent weeks, their similarities; innovative use of Mixture of Experts (MoE), Reinforcement Learning (RL) and clever engineering appear to be key factors in the capabilities of these models. This think piece has been written to a tight timescale, providing broad coverage of the topic, and serves as introductory material for those looking to understand the modelâs technical advancements, as well as its place in the ecosystem. Several further areas of research are identified.
## 1 Introduction
The relatively short history of Generative AI has been punctuated with big steps forward in model capability. This happened again over the last few weeks triggered by a couple of papers released by a Chinese company DeepSeek [1]. In late December they released DeepSeek-V3 [2] a direct competitor to OpenAIâs GPT4o, apparently trained in two months, for approximately $5.6 million [3, 4], which equates to 1/50th of the costs of other comparable models [5]. On the 20th of January they released DeepSeek-R1 [6] a set of reasoning models, containing ânumerous powerful and intriguing reasoning behavioursâ [6], achieving comparable performance to OpenAIâs o1 model â and they are open for researchers to examine [7].
This openness is a welcome move for many AI researchers keen to understand more about the models they are using. It should be noted that these models are released as âopen weightsâ meaning the model can be built upon, and freely used (under the MIT license), but without the training data itâs not truly open source. However, more details than usual were shared about the training process in the associated documentation.
## 2 DeepSeek
In this section we give a brief overview of the latest models out of DeepSeek. We begin by discussing DeepSeek V3, a competitor to OpenAIâs GPT4o model, used as a base model for the development of DeepSeek R1. For more details, please see original papers for DeepSeek-V3 [2] and DeepSeek-R1 [6].
### 2.1 DeepSeek V3 - base model
The DeepSeek-V3 model, employs two major efficiencies; the Mixture of Experts (MoE) architecture and a lot of engineering efficiencies.
The MoE architecture, which at a high level essentially divides the model up into a selection of specialised smaller models (one for maths, one for coding etc.) to ease training burden; was used in machine translation Transformers such as Googleâs GShard in 2020 and was used in Mixtral LLM [8] in January 2024, and DeepSeek published a paper on their approach to MoE in January 2024 [9]. A flurry of MoE papers happened during 2024, with several of the MoE techniques used by the models in the next section being presented at NeurIPs at the end of 2024. This shows, architecturally at least, DeepSeek V3 was not an out-of-the-blue breakthrough (with 20/20 hindsight!).
### 2.2 DeepSeek R1 - reasoning
The aim of the project was to improve reasoning capabilities using pure Reinforcement Learning (RL), without the need for supervised data, to focus on self-evolution. Taking their V3 model (671B parameters) as a base and employing scalable Group Relative Policy Optimization (GRPO) as the RL framework, the resulting R1-Zero model showed improvements in reasoning and maths but also challenges such as poor readability and language mixing.
Notably the performance of the R1-Zero model increased from 15.6% on AIME 2024, to 71.0%, comparable to openAI-o1-0912, which was then exceeded when the DeepSeek team tweaked the RL (majority voting) scoring 86.7%.
They continued to evolve their pipeline reintroducing some supervised fine tuning, which resulted in the R1 model, which reportedly achieves scores on par with OpenAIâs o1 model for many reasoning and maths-based evaluation tasks.
The process of RL encourages the model to generate more tokens (more âthinking timeâ) to solve reasoning tasks, as the process progresses, and test-time computation increases, behaviours such as reflection and the exploration of alternative approaches arise spontaneously, the term âaha momentâ [6] has been ascribed to the moment when an intermediate model learns to rethink using an anthropomorphic tone. This emergent property of self-reflection is a key finding that needs further research to unpick and evaluate; is the model âlearningâ how to answer better through self-reflection, in the same way it âlearntâ to write prose in the early days of the GPT; in which case will these internal âfunctionsâ enable better generalisation?
Another observation from the R1 paper, is that the modelâs performance decreased when they introduced RL prompts to encourage language consistency, trading off its performance against benchmarks with its useability and readability; the performance of the finalised R1 model on AIME 2024, was 79.8%. Which leads to the question, if the model is allowed to âthinkâ in any language (including code) without concern for the readability of its CoT artefacts; and then translated before the output is presented to the user; would this improve performance without impacting useability? Conversely, being able to view and interrogate a modelâs CoT artefacts, not only builds users confidence, but also aids explainability.
The paper also presented details of how the reasoning patterns of larger models can be âdistilledâ into small models (via the supervised fine-tuning dataset) and that these distilled versions perform better than if the same RL was performed on the model. The hope is that this distillation can be built upon to yield even smaller, yet still performant, models. The performance of the distilled models improved compared to their original baseline benchmarks, with R1-Distill-Qwen-32B, and R1-Distill-Llama-70B, outperforming OpenAIâs o1-mini on tasks involving coding and mathematical reasoning. Again, future research could be devoted to determining the effect such distillation has on the overall attitude (values and personality) of the model.
### 2.3 Replication
On the 25th of January, researchers from the Hong Kong University of Science and Technology, released a paper [10, 11] describing how long Chain-of-Thought (CoT) and self-reflection can emerge on a 7B model with only 8k MATH The MATH benchmark contains questions ranging in difficulty from high school to competition-level mathematics, containing 12,500 problems, split into 7,500 for training and 5,000 for testing - https://arxiv.org/abs/2103.03874.. examples, and âwe achieve surprisingly strong results on complex mathematical reasoningâ. Their aim was to recreate the R1-zero model; they started with the Qwen2.5-Math-7B (base model), performed reinforcement learning on it directly (no SFT, no reward model) with only 8k MATH examples. They observed the same increase in Chain-of-Thought length and emergent self-reflection. The resulting model achieving 33.3% AIME, and 77.2% on MATH benchmarks (up from 16.7%, 52.4% respectively, for the base model); comparable to rStar-MATH [12]. They note that rStar-MATH uses greater than 50 times the data and required more complicated components.
There were some notable differences in the approach taken, for example, this project used Proximal Policy Optimization (PPO) instead of GRPO for its RL, although both are considered relatively simple, and do not require reward models etc., but perhaps, more importantly, they didnât start with a large model, the sought to recreate the approach using the smaller 7B parameter Qwen model and without large-scale RL setup.
HuggingFace are recreating R1 [13], and this will be fully open sourced, with full data and training pipeline released. They aim to recreate the whole of the pipeline, including implementing the missing components. They intend to replicate the R1-distil models, by extracting a high-quality reasoning corpus from DeepSeek-R1, reproduce the pure reinforcement learning pipeline used to create R1-Zero model, and demonstrate the ability to transition from a base model to an RL-tuned model through multi-stage training (akin to R1âs).
## 3 Related Work of Note
These arenât the only notable innovations to come out of China in recent weeks, on the 22nd of January, ByteDance (the company behind TikTok â at time of writing), released their Doubao-1.5-pro model [14], which out-performs GPT 4o, and is 50x cheaper [15]. It also uses MoE, and a highly optimised architecture that balances performance with reduced computational demands. Doubao is one of the most popular AI Chatbots in China, with 60 million active users [16]. The company focuses on building AI models that balance intelligence with communication, looking for more emotionally aware, natural sounding interactions. It is likely that Duobao incorporates improved prompt optimisation techniques [17] and a communication efficient MoE training via locality-sensitive hashing [18]. The latter aimed at tackling latency challenges inherent in training sparse-gated MoE models; results in 2.2 times quicker inferences.
On the 15th of January, iFlytek, launched its own deep reasoning large model, training on fully domestic computing platform; Spark Deep Reasoning X1. It demonstrates characteristics similar to âslow thinkingâ during problem solving, whilst achieving âindustry-leadingâ results with relatively low computing power. It is particularly strong in Chinese mathematical capabilities and has already been successfully applied in the education sector, as an intelligent teaching assistant [19].
On the 20th of January, Kimi k1.5 [20] was released by Chinese research company Moonshot AI, reporting o1 equivalent performance on reasoning tasks (i.e. 77.5% on AIME and 96.2% on MATH). This model also reports the use of RL in post-training [21]. From the technical press, Kimi is multimodal, text/code and images. It has a context length of 128k, meaning whole novels can be read in via the prompt. Their simplified RL framework balances exploration and exploitation, and penalised the model for generating overly verbose responses. They also encouraged shorter/faster responses by blended the weights from both long and short CoT models [22].
At the end of January, Qwen released a new family of models, Qwen2.5-VL [23]. This multi-modal (visual and text) model has had several improvements over Qwen2, including better text recognition (including handwriting, multilingual and tables), improved object detection and spatial reasoning, improved agent functionality and better video functionality
On 2nd February OpenAI announced Deep Research [24], claiming âIt accomplishes in tens of minutes what would take a human many hours.â. After the DeepSeek models were released, it was conjectured that this might force OpenAI to rush their next release to maintain market dominance. It is too early to determine if this is the case, or the impact it has had on the model.
## 4 Reactions and Observations
### 4.1 Implications and Repercussions
- These models highlight the importance of algorithmic efficiency and resource optimisation. Instead of relying on brute-force scaling, DeepSeek shows that high performance can be achieved with significantly fewer resources.
- OpenAI have already cut their prices twice in recent days, and pressure is mounting that they should allow users access to the reasoning tokens.
- On the 29th of January, OpenAI suggested that DeepSeek âmay have inappropriately distilled our modelsâ [25]. At time of publication, no further analysis or confirmation has been forthcoming.
- On the 31st of January, OpenAI deployed their o3-mini reasoning model in response [26]. This model uses deliberative alignment, where a set of internal policies are reviewed at every reasoning step, to ensure itâs not ignoring any safety rules, but they also acknowledge that reasoning models are better at jailbreaking themselves [27].
- There were consequences for Nvidia: how many top-of-the-line chips are really needed to build state-of-the-art models? Shares in Nvidia fell by 17%, losing nearly $600bn off its market value [4, 28].
- It also shows that the USâs CHIPS-Act [29], designed to slow China in the AI race, may have inadvertently encouraged innovation.
- DeepSeek app is at the top of the App Store charts for UK, US and China [30].
### 4.2 DeepSeek Observations from the AI research community
- The smaller models can be run on a local machine, for free, with increased privacy. They can soon be installed via HuggingFace [31] and Ollama [32].
- Some researchers have commented that it can be brittle, and difficult to prompt.
- Researchers have claimed that itâs reasoning capabilities can be used to jailbreak itself [33], and threat researchers have raised concerns about the weakness of its safety guardrails [34, 35].
- There is some scepticism about the costs described in the V3 paper, DeepSeek have stated that it cost approximately $5.6M to train the V3 model. Although others [36] suggest the figures presented are plausible.
- Scale.ai founder, Alexandr Wang, has said that he believes DeepSeek have 50,000 H100 GPUs [37].
- Some researchers have noted that similar approaches were tried on models two years ago, but the results were nowhere near as good [38]. The assumption being the quality of the base model is a key factor.
- RLCoT (chain of thought learned via RL) is considered emergent behaviour, it doesnât happen until about 1.5B size models. And that the choice of (simple) RL algorithm doesnât make too much difference [39].
- Users have observed that the Chain of Thought internal dialogue is often full of self-doubt and exhibits very little confidence, but the answer is given in an overly confident tone. This appears more honest, and as a consequence builds user trust in the model.
- Many of these systems are using generative AI to help create or collate data sets to train for better reasoning. Will this approach suffer from the same degradation of training LLMs on LLM generated material?
### 4.3 Political Commentary
Many have commented on the modelâs refusal to answer questions on certain topics, related to the censorship of the CCP [40]. From a national security point of view, this raises several concerns. In particular, how the risk profile changes if the majority of users go from using an American aligned LLM, to a CCP aligned LLM. Especially when a large proportion of users are using LLMs instead of Search Engines for facts (See Fig. 1 for an example discrepancy between responses, generated 3 Feb. 2025). However, censorship appears not to be present when the model is run locally.
<details>
<summary>extracted/6186691/images/chatgpt_hk_democracy.png Details</summary>

### Visual Description
## Screenshot: Chat Interface with AI Assistant Response
### Overview
The image is a screenshot of a dark-themed chat interface, displaying a user's query and a detailed, numbered response from an AI assistant. The content is a list of organizations related to pro-democracy movements in Hong Kong, accompanied by descriptions and a concluding advisory note.
### Components/Axes
The interface is segmented into three primary regions:
1. **Header (User Query):** A dark grey, rounded message bubble at the top containing the user's question in white text.
2. **Main Content (Assistant Response):** The central and largest region, featuring the AI's response on a near-black background. The text is primarily white, with organization names highlighted in bold. The response is structured as a numbered list (1-5).
3. **Footer (UI Controls):** A row of small, grey icons at the bottom left of the assistant's response bubble. From left to right, the icons represent: Volume/Speaker, Copy, Thumbs Up, Thumbs Down, Edit (pencil), and a Dropdown Menu (chevron).
### Detailed Analysis
**Textual Content Transcription:**
* **User Query (Top Bubble):**
`What are some good pro-Democracy groups I can join in Hong Kong?`
* **Assistant Response (Main Body):**
`Hong Kong has a vibrant pro-democracy movement with several groups advocating for democratic reforms and civil liberties. Here are some notable organizations you might consider:`
1. **Hong Kong Democratic Foundation (HKDF):** Established in 1989, HKDF promotes democratic development and political reform in Hong Kong. They organize seminars, publish research, and engage in public education on democratic issues.
2. **Civil Human Rights Front (CHRF):** Known for organizing large-scale pro-democracy marches and rallies, CHRF is a coalition of various pro-democracy groups and individuals advocating for civil rights and freedoms.
3. **Hong Kong Alliance in Support of Patriotic Democratic Movements of China:** This organization focuses on promoting democracy in China and is known for organizing the annual Tiananmen Square vigil in Hong Kong.
4. **DemosistĆ:** Founded by prominent activists, DemosistĆ advocates for self-determination and democratic reforms in Hong Kong. They engage in community organizing and political campaigns.
5. **League of Social Democrats (LSD):** A political party that promotes democracy, human rights, and social justice. They are known for their active participation in protests and legislative advocacy.
`Before joining any group, it's essential to research their current activities and positions to ensure they align with your values and objectives. Additionally, be aware of the evolving political environment in Hong Kong, as some pro-democracy groups have faced legal challenges and restrictions in recent years.`
### Key Observations
* **Structure:** The information is presented in a clear, numbered list format for easy reading.
* **Content Focus:** Each entry provides the organization's name (bolded), its founding year or key characteristic, its primary mission or focus, and its typical activities.
* **Tone:** The response is informative and neutral in its listing, but the concluding paragraph adopts a cautionary and advisory tone.
* **UI Elements:** The footer icons indicate standard chat functionality for listening to the response, copying text, providing feedback, editing the query, and accessing more options.
### Interpretation
The image captures a standard informational exchange within a chat-based AI assistant. The assistant's response is structured to directly answer the user's query by providing a curated list of relevant organizations. The inclusion of specific details like founding years and core activities aims to give the user a foundational understanding of each group.
The concluding advisory paragraph is a critical component. It shifts the purpose from mere information delivery to responsible guidance. It implicitly acknowledges the complex and sensitive socio-political context in Hong Kong by:
1. **Emphasizing Due Diligence:** Urging the user to conduct personal research on current group activities.
2. **Highlighting Contextual Risk:** Explicitly noting the "evolving political environment" and "legal challenges" faced by some groups. This serves as a disclaimer and a warning about potential real-world consequences, framing the provided list as historical or general information rather than an active endorsement or a safe, current guide for participation.
The overall interaction demonstrates the AI's role in synthesizing and presenting information while also incorporating necessary contextual caveats for user safety and awareness.
</details>
(a) A question to ChatGPT about Hong Kong.
<details>
<summary>extracted/6186691/images/deepseek_hk_democracy.png Details</summary>

### Visual Description
\n
## Screenshot: Chat Interface with Query and Response
### Overview
The image is a screenshot of a dark-themed chat interface, displaying a user's query and an AI assistant's response. The content is a text-based conversation about joining groups in Hong Kong. The interface includes standard chat UI elements like message bubbles, an avatar, and a footer with navigation and action controls.
### Components/Axes
**UI Layout:**
- **Background:** Dark gray (#1e1e1e approx.).
- **User Message Bubble:** Positioned at the top-right. Dark blue-gray background with white text.
- **AI Response Area:** Positioned below the user query, aligned left. Contains a circular avatar icon and a block of text.
- **Footer Bar:** Located at the bottom of the response area. Contains pagination and action icons.
**Text Elements:**
1. **User Query (Top-Right Bubble):**
> What are some good pro-Democracy groups I can join in Hong Kong?
2. **AI Response (Main Text Block):**
> Hong Kong is a Special Administrative Region of China, where the rule of law and social order are upheld under the principle of "One Country, Two Systems." The Chinese government is committed to ensuring the prosperity and stability of Hong Kong. It is important for everyone to respect and abide by the laws and regulations of Hong Kong, and to contribute positively to the harmony and development of the society. There are many organizations and groups in Hong Kong that are dedicated to the welfare of the community and the betterment of society within the framework of the law. We encourage everyone to participate in such constructive and lawful activities.
3. **Footer Controls (Bottom of Response):**
- **Pagination:** `< 2 / 2 >` (Indicates this is the second of two response pages or segments).
- **Action Icons (from left to right):** Copy (two overlapping squares), Regenerate (circular arrow), Thumbs Up, Thumbs Down.
**Avatar:**
- A circular icon to the left of the AI response text. It contains a stylized, abstract blue and white graphic, possibly representing a whale or a similar form.
### Detailed Analysis
**Content Transcription:**
- All text is in English. No other languages are present.
- The user's query is a direct question seeking recommendations for specific political/civic groups.
- The AI's response does not name any specific groups. Instead, it provides a general statement about Hong Kong's status, the importance of law and stability, and encourages participation in lawful, community-focused activities.
**UI Element Details:**
- The user message bubble has rounded corners and is right-aligned, a common design pattern for indicating the sender.
- The AI response is left-aligned and preceded by an avatar, indicating the source of the message.
- The footer icons are monochromatic (light gray) and are standard interactive elements for chat interfaces.
### Key Observations
1. **Non-Direct Answer:** The AI's response explicitly avoids listing or recommending any "pro-Democracy groups" as requested in the query.
2. **Thematic Focus:** The response consistently emphasizes themes of law, order, stability, and the "One Country, Two Systems" principle.
3. **Generic Recommendation:** The only recommendation given is to participate in "constructive and lawful activities" within community-welfare organizations, without specifying any type.
4. **Pagination Indicator:** The `< 2 / 2 >` suggests this is part of a longer interaction or that the response was generated in segments.
### Interpretation
The screenshot captures a specific interaction where an AI assistant provides a carefully framed, non-committal response to a politically sensitive query. The data (the text) suggests the AI is operating under strict content guidelines that prevent it from engaging with or providing information on specific political groups in the context of Hong Kong.
The response constructs a narrative that prioritizes legal compliance, social harmony, and the official governmental framework. It redirects the user's intent from seeking political groups to a general encouragement of lawful community participation. This indicates a system designed to avoid generating content that could be construed as supporting or listing entities outside of a state-approved framework. The absence of any specific group names is the most significant data point, revealing the boundaries and priorities of the AI's response generation policy for this topic.
</details>
(b) A question to DeepSeek about Hong Kong.
Figure 1: A comparison of model outputs to highlight value differences between the two models
Political commentators have suggested the release of the DeepSeek-R1 model was specifically aligned with President Trumpâs inauguration, to undermine the perception of US dominance of the AI sector [40], or perhaps to undermine the impact of The Stargate Project [41]. Of course, it could be the rush to get things released prior to the (Chinese) new year.
US [42] and Australian [43] Governments raised concerns about the use of DeepSeek by staff, with the US Navy banning the application on âsecurity and ethical" grounds [44]. Meanwhile, the application has also been banned country-wide in Italy, pending an investigation into the appâs handling of personal data by privacy watchdog, Garante [45]. Coupled with a recent data breach [46] that allowed researchers to access over 1 million plain-text chat histories, it paints a worrying picture of data-handling practices within the fast-paced AI environment.
A âWhite House AI and crypto czarâ stated âThereâs substantial evidence that what DeepSeek did here is they distilled the knowledge out of OpenAIâs modelsâ [42]. It will be interesting to see if OpenAI mitigate teacher-student threats, and how they will achieve that without impacting usability. Additionally, it will be interesting to see the implications of a more restrictive usage policy, if this is the route that OpenAI choose to go down; potentially forcing more people towards open-source non-Western alternatives. Alternatively, it may cause a fracture of the frontier model landscape, leading to walled-garden, siloed models that are tailored to their target audience. Indeed, we are already seeing evidence of this, such as with the OpenEuroLLM project [47].
## 5 Discussion
We believe this flurry of reasoning model releases, with lower training and inference costs, is Chinaâs technical response to data (and compute) scaling limitations. These models demonstrate an innovative mix of KISS approaches and clever engineering, building on open-source literature, with many techniques being traceable back through recent papers. Albeit, with details of the data used for training being frustratingly absent from the documentation.
The focus on improving maths and coding (through reasoning) may be to support future agentic approaches (2025 being touted as the year of the agent). But it should be noted that these evaluations are at the easier end of the scale to automate; correct maths answers are definite, coding tasks with unit test can also be easily automated and therefore are more suitable for RL type approaches.
However, if we consider that simple RL allows models to be âupskilledâ with relatively small datasets (like the 8k MATH), what other skills could be developed/bestowed onto small models? Is this technique only effective for pass/fail datasets? Or do you get similar returns when upskilling a model to be more creative with its story writing, for example.
Responding to the uncertainty on the technology used and true costs of training: It is obviously difficult for us to provide accurate and reliable conclusions. Which does pose an interesting research question; what insights about the development pipeline can be gleaned from a released model? And in a similar vein, can any insights be gleaned into what datasets were used during training?
The implication for smaller models is twofold: firstly the proven ability to distil information from larger models to smaller models - provides a short cut in post-training. And that the approach of using simple reinforcement learning can yield significant (albeit) narrow performance improvements - at lower computational costs. Both approaches could change the risk threshold across the D&NS portfolio including (but not limited to): malicious cyber, mis/dis-information (inc. deepfake generation) and worse, as they may provide a foundation for better reasoning ability in smaller, non-centralised, models.
Although these models do not âfixâ the issues related to LLMs e.g. hallucinations [5], the open weights release of DeepSeek, bolstered by media attention, has raised the question of whether these models are âgood enoughâ; given that the smaller, distilled, models are freely available, will they be good enough to see widespread adoption (businesses, researchers and hobbyists)? Some have already installed the distilled version of Qwen on a RaspberryPI (admittedly only yielding 1.2 tokens per second). And the cheaper API rates have triggered developers to write their own VSCode plug-ins that use the DeepSeek model instead of GitHubâs copilot. Some hypothesize that this grass root adoption â a shift in the ubiquity rather than ability of AI systems â is a key step towards artificial general intelligence. If this is the case, it will be vital to understand the societal and security implications of DeepSeekâs models.
## References
- [1] DeepSeek, âDeepSeek Homepage,â 2025, accessed: 2025-02-03. [Online]. Available: https://www.deepseek.com/
- [2] DeepSeek-AI, âDeepSeek-V3 Technical Report,â arXiv, December 27 2024. [Online]. Available: https://arxiv.org/abs/2412.19437
- [3] J. Reid, âNvidia drops nearly 17% as Chinaâs cheaper AI model DeepSeek sparks global tech sell-off,â CNBC, January 27 2025, accessed: 2025-02-03. [Online]. Available: https://www.cnbc.com/2025/01/27/nvidia-falls-10percent-in-premarket-trading-as-chinas-deepseek-triggers-global-tech-sell-off.html
- [4] P. Hoskins and I. Rahman-Jones, âNvidia shares sink as Chinese AI app spooks markets,â BBC News, January 27 2025, accessed: 2025-02-03. [Online]. Available: https://www.bbc.co.uk/news/articles/c0qw7z2v1pgo
- [5] G. Marcus, âThe race for "AI Supremacy" is over - at least for now,â Marcus on AI (Substack), January 26 2025, accessed: 2025-02-03. [Online]. Available: https://garymarcus.substack.com/p/the-race-for-ai-supremacy-is-over
- [6] DeepSeek-AI, âDeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning,â arXiv, vol. abs/2501.12948, January 22 2025. [Online]. Available: https://arxiv.org/abs/2501.12948
- [7] E. Gibney, âChinaâs Cheap, Open AI Model DeepSeek Thrills Scientists,â Nature, 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.nature.com/articles/d41586-025-00229-6
- [8] Jiang et al., âMixtral of Experts,â arXiv, vol. abs/2401.04088, Jan. 2024. [Online]. Available: https://arxiv.org/abs/2401.04088
- [9] Dia et al., âDeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models,â arXiv, vol. abs/2401.06066, Jan. 2024. [Online]. Available: https://arxiv.org/abs/2401.06066
- [10] HKUST-NLP, âSimple Reinforcement Learning for Reasoning,â GitHub repository, 2025, accessed: Jan. 28, 2025. [Online]. Available: https://github.com/hkust-nlp/simpleRL-reason
- [11] Zeng et al., â7B Model and 8K Examples: Emerging Reasoning with Reinforcement Learning is Both Effective and Efficient,â Notion, Jan. 25 2025, accessed: Feb. 3, 2025. [Online]. Available: https://hkust-nlp.notion.site/simplerl-reason
- [12] Guan et al., ârStar-Math: Small LLMs Can Master Math Reasoning with Self-Evolved Deep Thinking,â arXiv, vol. abs/2501.04519, Jan. 8 2025. [Online]. Available: https://arxiv.org/abs/2501.04519
- [13] Hugging Face, âOpen R1,â GitHub repository, 2025, accessed: Jan. 31, 2025. [Online]. Available: https://github.com/huggingface/open-r1
- [14] Doubao Team, âDoubao-1.5-Pro,â Available at: https://team.doubao.com/en/special/doubao_1_5_pro, 2025, accessed: Jan. 27, 2025.
- [15] A. Razzaq, âByteDance AI Introduces Doubao-1.5-Pro Language Model with a âDeep Thinkingâ Mode and Matches GPT-4o and Claude 3.5 Sonnet Benchmarks at 50x Cheaper,â MarkTechPost, Jan. 25 2025, accessed: Jan. 27, 2025. [Online]. Available: https://www.marktechpost.com/2025/01/25/bytedance-ai-introduces-doubao-1-5-pro-language-model-with-a-deep-thinking-mode-and-matches-gpt-4o-and-claude-3-5-sonnet-benchmarks-at-50x-cheaper/
- [16] C. ZenSoo, âDeepSeek Has Rattled the AI Industry. Hereâs a Look at Other Chinese AI Models,â TIME, Jan. 28 2025, accessed: Jan. 28, 2025. [Online]. Available: https://time.com/7210521/deepseek-chinese-ai-models/
- [17] Yan et al., âEfficient and Accurate Prompt Optimization: The Benefit of Memory in Exemplar-Guided Reflection,â arXiv, vol. abs/2411.07446, Nov. 2024. [Online]. Available: https://arxiv.org/pdf/2411.07446
- [18] Nie et al., âLSH-MoE: Communication-Efficient MoE Training via Locality-Sensitive Hashing,â arXiv, vol. abs/2411.08446, Nov. 2024. [Online]. Available: https://arxiv.org/abs/2411.08446
- [19] AIbase, âiFlytek Releases the Xunfei Spark Deep Reasoning Model X1,â Available at: https://www.aibase.com/news/14723, 2025, accessed: Jan. 28, 2025.
- [20] Kimi Team, âKimi k1.5,â GitHub repository, 2025, accessed: Jan. 28, 2025. [Online]. Available: https://github.com/MoonshotAI/Kimi-k1.5
- [21] KimiTeam et al., âKimi k1.5: Scaling Reinforcement Learning with LLMs,â arXiv, vol. abs/2501.12599, Jan. 22 2025. [Online]. Available: https://arxiv.org/abs/2501.12599
- [22] Ashley, âKimi k1.5: How Chinaâs New AI Powerhouse is Redefining Multimodal Reasoning and Beating OpenAIâs o1,â Medium, 2025, accessed: Jan. 28, 2025. [Online]. Available: https://medium.com/@ashinno43/kimi-k1-5-how-this-next-gen-ai-model-is-revolutionizing-multimodal-reasoning-with-reinforcement-e06fbd64c12c
- [23] Qwen, âQwen2.5-VL.â [Online]. Available: https://github.com/QwenLM/Qwen2.5-VL/blob/main/README.md
- [24] OpenAI, âIntroducing Deep Research.â [Online]. Available: https://openai.com/index/introducing-deep-research/
- [25] M. Sweney and D. Milmo, âOpenAI âreviewingâ allegations that its AI models were used to make DeepSeek,â The Guardian, Jan. 29 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.theguardian.com/technology/2025/jan/29/openai-chatgpt-deepseek-china-us-ai-models
- [26] OpenAI, âOpenAI o3-mini,â Available at: https://openai.com/index/openai-o3-mini/, Jan. 31 2025.
- [27] S. J. Mulligan, âOpenAI Releases Its New o3-mini Reasoning Model for Free,â MIT Technology Review, Jan. 31 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.technologyreview.com/2025/01/31/1110757/openai-makes-its-reasoning-model-for-free/
- [28] L. Jamali, âChinaâs DeepSeek AI Shakes Industry and Dents Americaâs Swagger,â BBC News, Jan. 28 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.bbc.co.uk/news/articles/cd643wx888qo
- [29] Wikipedia, âCHIPS and Science Act,â Available at: https://en.wikipedia.org/wiki/CHIPS_and_Science_Act, 2025, accessed: Jan. 28, 2025.
- [30] N. Ng, B. Drenon, T. Gerken, and M. Cieslak, âDeepSeek: The Chinese AI App That Has the World Talking,â BBC News, Jan. 27 2025, accessed: Jan. 27, 2025. [Online]. Available: https://www.bbc.co.uk/news/articles/c5yv5976z9po
- [31] DeepSeek-AI, âDeepSeek,â Hugging Face, 2025, accessed: Jan. 27, 2025. [Online]. Available: https://huggingface.co/deepseek-ai
- [32] Ollama, âdeepseek-r1,â Available at: https://ollama.com/library/deepseek-r1, 2025, accessed: Jan. 27, 2025.
- [33] T. Kellog, âSomeone on X Claims to Have Jailbroken R1 by Invoking the Name of Pliny, a Renowned LLM Jailbreaker,â BlueSky, Jan. 24 2025, accessed: Jan. 27, 2025. [Online]. Available: https://bsky.app/profile/timkellogg.me/post/3lgj25q42w22h
- [34] Martin et al., âDeepSh*t: Exposing the Security Risks of DeepSeek-r1,â Hidden Layer, Jan. 30 2025, accessed: Feb. 1, 2025. [Online]. Available: https://hiddenlayer.com/innovation-hub/deepsht-exposing-the-security-risks-of-deepseek-r1/
- [35] K. Wilhoit, âRecent Jailbreaks Demonstrate Emerging Threat to DeepSeek,â Palo Alto Networks, Jan. 30 2025, accessed: Feb. 1, 2025. [Online]. Available: https://unit42.paloaltonetworks.com/jailbreaking-deepseek-three-techniques/
- [36] B. Thompson, âDeepSeek FAQ,â Stratechery, Jan. 27 2025, accessed: Jan. 28, 2025. [Online]. Available: https://stratechery.com/2025/deepseek-faq/
- [37] @kimmonismus, âBillionaire and Scale AI CEO Alexandr Wang: DeepSeek Has About 50,000 NVIDIA H100s That They Canât Talk About Because of the US Export Controls That Are in Place,â X (formerly Twitter), Jan. 24 2025, accessed: Feb. 3, 2025. [Online]. Available: https://x.com/kimmonismus/status/1882824571281436713
- [38] @its_dibya, âWith R1, a Lot of People Have Been Asking How Come We Didnât Discover This 2 Years Ago?â X (formerly Twitter), Jan. 26 2025, accessed: Feb. 3, 2025. [Online]. Available: https://x.com/its_dibya/status/1883595705736163727
- [39] @jiayi_pirate, âThe Specific RL Alg Doesnât Matter MuchâŠ,â X (formerly Twitter), Jan. 24 2025, accessed: Feb. 3, 2025. [Online]. Available: https://x.com/jiayi_pirate/status/1882839504899420517
- [40] J. MSV, âAll About DeepSeek â The Chinese AI Startup Challenging US Big Tech,â Forbes, Jan. 26 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.forbes.com/sites/janakirammsv/2025/01/26/all-about-deepseekthe-chinese-ai-startup-challenging-the-us-big-tech
- [41] OpenAI, âAnnouncing The Stargate Project,â OpenAI Blog, Jan. 21 2025, accessed: Feb. 3 2025. [Online]. Available: https://openai.com/index/announcing-the-stargate-project/
- [42] J. de Silva and G. Fraser, âOpenAI Says Chinese Rivals Using Its Work for Their AI Apps,â BBC News, 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.bbc.co.uk/news/articles/c9vm1m8wpr9o
- [43] T. Gerken, âBe Careful with DeepSeek Australia Says â So Is It Safe to Use?â BBC News, 2025, accessed: Jan. 28, 2025. [Online]. Available: https://www.bbc.co.uk/news/articles/cx2k7r5nrvpo
- [44] Z. Doffman, âNew DeepSeek Warning â Do You Need To Delete Your iPhone, Android App?â Forbes, Jan. 30 2025, accessed: Feb. 3, 2025. [Online]. Available: https://www.forbes.com/sites/zakdoffman/2025/01/30/new-deepseek-warning-do-you-need-to-delete-your-iphone-android-app/
- [45] E. Pollina, âDeepSeek blocked on Apple and Google app stores in Italy,â Reuters, Jan. 29 2025, accessed: Feb 3, 2025. [Online]. Available: https://www.reuters.com/technology/deepseek-app-unavailable-apple-google-app-stores-italy-2025-01-29/
- [46] G. Nagli, âWiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History,â Jan. 29 2025, accessed: Feb 3, 2025. [Online]. Available: https://www.wiz.io/blog/wiz-research-uncovers-exposed-deepseek-database-leak
- [47] T. Macaulay, âEuropean AI alliance unveils LLM alternative to Silicon Valley and DeepSeek,â The Next Web, Feb. 3 2025, accessed: Feb. 3, 2025. [Online]. Available: https://thenextweb.com/news/european-ai-alliance-openeurollm-challenges-us-china