DeepSeek‘s AI: Did It Learn From Google’s Gemini?
The AI community is abuzz with speculation that Chinese startup DeepSeek may have trained its latest model, R1-0528, using outputs from Google’s Gemini. While unconfirmed, this possibility raises important questions about AI training methodologies and the use of existing models.
Traces of Gemini in DeepSeek‘s R1-0528
AI researcher Sam Paech observed that DeepSeek‘s R1-0528 exhibits linguistic patterns and terminology similar to Google’s Gemini 2.5 Pro. Terms like “context window,” “foundation model,” and “function calling”—commonly associated with Gemini—appear frequently in R1-0528’s outputs. These similarities suggest that DeepSeek may have employed a technique known as “distillation,” where outputs from one AI model are used to train another. linkedin.com
Ethical and Legal Implications
Using outputs from proprietary models like Gemini for training purposes raises ethical and legal concerns. Such practices may violate the terms of service of the original providers. Previously, DeepSeek faced similar allegations involving OpenAI‘s ChatGPT. androidheadlines.com

DeepSeek‘s R1-0528: Performance and Accessibility
Despite the controversy, R1-0528 has demonstrated impressive performance, achieving near parity with leading models like OpenAI‘s o3 and Google’s Gemini 2.5 Pro on various benchmarks. The model is available under the permissive MIT License, allowing for commercial use and customization.
As the AI landscape evolves, the methods by which models are trained and the sources of their training data will continue to be scrutinized. This situation underscores the need for clear guidelines and ethical standards in AI development.
For more information, you can refer to the following articles:
- DeepSeek may have used Google’s Gemini to train its latest model
- DeepSeek’s Secret Sauce? A Dash of Google Gemini
- DeepSeek R1-0528 arrives in powerful open source challenge to OpenAI o3 and Google Gemini 2.5 Pro
Exploring the Possibility
The possibility of DeepSeek utilizing Google’s Gemini highlights the increasing interconnectedness of the AI landscape. Companies often use pre–trained models as a starting point and fine-tune them for specific tasks. This process of transfer learning can significantly reduce the time and resources required to develop new AI applications. Understanding transfer learning and its capabilities is important when adopting AI tools and platforms. DeepSeek might have employed a similar strategy.
Ethical Implications and Data Usage
If DeepSeek did, in fact, use Gemini, it brings up some ethical concerns. Consider these factors:
- Transparency: Is it ethical to use a competitor’s model without clear acknowledgment?
- Data Rights: Did DeepSeek have the right to use Gemini’s outputs for training?
- Model Ownership: Who owns the resulting AI model, and who is responsible for its outputs?
These are critical questions within the AI Ethics and Impact space and need careful consideration as AI technology advances. The use of data from various sources necessitates a strong understanding of data governance. You can learn more on data governance using Oracle data governance.
DeepSeek‘s Response
As of now, DeepSeek hasn’t officially commented on these rumors. An official statement from DeepSeek would clarify the situation. A response would help us understand their development process and address any ethical concerns.