Site icon Digital Creed

DeepSeek: The Whale in the Deep Blue Sea

DeepSeek AI prompt phone screen

Image: Solen Feyissa on Unsplash

When DeepSeek’s R1 model surpassed OpenAI’s ChatGPT on the app charts, it drew global attention and raised data privacy concerns. How has DeepSeek-R1 impacted the big tech companies, and how did they respond? What are the lessons learned?

Image: Solen Feyissa on Unsplash

First, an explanation for this story headline. The Whale is DeepSeek’s logo. The Deep Blue Sea is a metaphor for deep learning and reasoning models, which have garnered much interest lately.

For the uninitiated, DeepSeek R1 is a family of AI models based on reinforcement learning (RL) that’s designed for logical and reasoning tasks. The model solves complex problems by breaking them down into multiple steps. It’s open-source (with the MIT license) and has a conversational chat interface like any other genAI tool.

Reinforcement learning is a term that stands out in that definition. In its simplest form, AI models now think twice or more before spitting out answers to user prompts. In fact, you will now see “thinking….” In the interface of RL models. Remember how your high school teacher or mama told you, “Think before you speak”?

According to Geeks-for-Geeks, Reinforcement Learning is a branch of machine learning focused on making decisions to maximize cumulative rewards in a given situation. Unlike supervised learning, which relies on a training dataset with predefined answers, RL involves learning through experience. In RL, an agent learns to achieve a goal in an uncertain, potentially complex environment by performing actions and receiving feedback through rewards or penalties.

That means you are going to get more accurate answers. The models use techniques such as backpropagation and weights in neural networks to evaluate the most accurate and relevant response to a prompt.

Lessons Learned

Now, back to the Chinese company DeepSeek and its models.

The biggest lesson we learned from DeepSeek is that you don’t need cutting-edge technology (expensive GPUs) to train a model. DeepSeek developed and trained its R1 model for just $6 million, whereas it cost OpenAI $500 million to train its O1 model. Also, due to trade sanctions imposed by the U.S. on Chinese companies, DeepSeek did not have access to the latest hardware – it used Nvidia GPUs purchased in 2023.

This shift highlights how countries can get around the challenge of accessing the latest technology from Western countries due to trade sanctions (chip export controls) and high tariffs.

Who’s Using DeepSeek Models?

After the hue and cry about data privacy and concerns about data stored on Chinese servers, companies are now taking DeepSeek seriously.

I have yet to hear about user companies using DeepSeek models, which is going to take a while, due to regulatory and compliance concerns. Governments are rushing to ban departments from using it citing security and privacy concerns – primarily sensitive national data leaking to the Chinese government.

However, the product and service companies are integrating DeepSeek into their offerings. They are mostly taking the open-source route. The open-source version of DeepSeek R1 is available as an MIT license and this is the route these companies are taking. Since it is open-source, they can examine the code, tailor it to their specific needs and incorporate security and privacy controls to ensure data sovereignty.

Here are some examples:

Chinese Companies Too

Chinese companies are also integrating DeepSeek models:

Conclusion

It seems like some companies, especially startups, are willing to take the risk and adopt DeepSeek’s models – as they can drastically reduce the cost of model development and training. Yet this becomes a supply chain concern as startups usually innovate for large companies (banks and telecom companies, for instance), who, in turn, have customers. Large companies are custodians of their customer data, and they need to ensure that this data is safe in the hands of startups – or maybe just give them synthetic data to test their models, without taking that risk.

Exit mobile version