Investment Topics

DeepSeek's Success Boosts China's AI Sector

Advertisements

The rise of DeepSeek has garnered significant attention globally, not just for its innovative technology but also for the opportunities it has unlocked in the artificial intelligence domainSince its launch less than a month ago, DeepSeek has already emerged as the fastest-growing AI application worldwide, with a sharp increase in daily active usersAccording to statistics from AI product rankings, as of January 31, 2023, the global daily active users of DeepSeek have surpassed 20 million, surpassing ByteDance's Doubao product, while accounting for about 41.6% of ChatGPT's user base.

Despite this rapid ascent, users have often faced issues when trying to engage in frequent and deep conversations with DeepSeekInstances of server delays are common, frequently resulting in notifications stating, "The server is busy, please try again later." This has led some users to humorously speculate that DeepSeek's model being dubbed R1 is tied to its ability to "run" only once each day, underscoring the challenges of scaling such a promising technology.

On February 6, DeepSeek officially announced the temporary suspension of API service top-ups due to limited server resourcesAs of the time of writing, this recharge service remains unavailableAI professionals have noted that their teams had developed AI search functionalities based on the DeepSeek model, but following its explosive popularity, they experienced significant API service delays and response time-outs, which in turn stymied the generation of search resultsDuring the Spring Festival, they resorted to working overtime to shift services to a backup GPT-4o model.

In a larger context, the emergence of DeepSeek marks a "breakthrough" moment for AI technologies, presenting a multitude of promising opportunities for both upstream and downstream industriesCloud providers and chip manufacturers, armed with substantial computational capabilities, have rapidly mobilized in response.

Leading the charge were major cloud providers, both domestic and international

Advertisements

Notably, early in the Lunar New Year celebrations, tech giants like Microsoft and Amazon integrated the DeepSeek-R1 model into their cloud platformsFollowing suit, major Chinese cloud firms such as Huawei Cloud, Alibaba Cloud, Baidu Intelligent Cloud, ByteDance's Volcano Engine, and Tencent Cloud announced on February 1 the availability of the DeepSeek model for developers and corporate clients.

Simultaneously, a number of domestic chip manufacturers have proclaimed their successful adaptation and deployment of the DeepSeek modelCompanies like Mutex, Timesys Intelligence, Moer Thread, and Birun Technology have publicly stated that they have completed the necessary alignments with the DeepSeek modelThese chip manufacturers either utilized their own computational platforms or formed alliances with downstream AI infrastructure platforms to facilitate the model's deployment.

An industry insider explained that the swift action taken by these cloud providers can largely be attributed to the relatively low costs associated with integrating the DeepSeek modelThe model has been trained using NVIDIA GPUs, which many cloud providers have in abundance, allowing for direct and efficient deploymentIn contrast, domestic chip manufacturers, given their unique hardware instruction sets, face a more complex task of adaptation and migration, resulting in higher associated costs.

Whether they are cloud providers or chip manufacturers, there is a shared desire to capitalize on the buzz surrounding DeepSeekAs DeepSeek's official API service faced instability, many companies sought to redirect users to their platforms, leveraging their existing computational resources to offer DeepSeek model servicesUsers who explored these alternative platforms indicated that in many cases, the pricing and response times were adequate, prompting them to consider developing AI applications based on DeepSeek-R1 through third-party services.

Social media platforms also witnessed a surge of promotional materials from various third-party services, claiming to provide seamless usage experiences that circumvented the congestion on DeepSeek's official site

Advertisements

Some platforms even boasted of their integration of "domestic chips + local large models." For instance, Silicon Valley Flow collaborated with Huawei Cloud, deploying the DeepSeek model on their large model cloud service platformHuawei even embedded DeepSeek-R1 into the purely domestic Harmony version of its Xiao Yi assistant app.

Yuan Jinhui, the founder and CEO of Silicon Valley Flow, disclosed on social media that prior to the release of the DeepSeek-V3 model, founder Liang Wenfeng suggested deploying the model on their platform using least 20 NVIDIA H800 serversCiting cost concerns, they opted not to pursue that route initially.

Only after witnessing DeepSeek's meteoric success did the Silicon Valley Flow team decide to proceed with the adaptation of domestic chipsThis led them to partner with Huawei, and throughout the Spring Festival holiday, they worked diligently, discussing issues at all hours, even holding late-night meetingsUltimately, on February 1, they successfully launched their DeepSeek model service utilizing domestic chips.

The intersection of DeepSeek and domestic chips presents an intriguing narrativeDifferentiating between the stages of training and inference of a large model is critical to understanding how these two elements interplayDuring the training phase, a large model is constantly evolving, requiring extensive datasets and ongoing adjustments to internal parameters to recognize patternsInference, by contrast, represents the practical application of the model once it has completed its training.

A former AI engineer from a prominent tech firm elaborated that while the training phase demands higher computational power and bandwidth, it typically involves experimenting with various model architectures and operations, usually resulting in a preference for NVIDIA's hardware and its development toolkit, CUDAThe inference phase, having less stringent hardware and software requirements, provides a fertile ground for many domestic chips aiming to optimize and accommodate well-trained models.

Several domestic chip manufacturers asserted that while DeepSeek introduced minor innovations structurally, it remains a large language model

Advertisements

Adapting DeepSeek primarily occurs at the inference application stage, rendering it relatively straightforward and quick to implement.

The spotlight on DeepSeek’s cost-effectiveness resulted in dramatic fluctuations in the stock market, with NVIDIA's shares collectively dropping significantly, a record-breaking loss in market capitalization within a single dayOne prevalent theory circulating claimed that during its model development, DeepSeek circumvented the CUDA framework created by NVIDIA, thereby reducing dependency on NVIDIA processorsThis claim originated from the assertion in DeepSeek's V3 model technical report that they implemented tailored PTX (Parallel Thread Execution) instructions while automatically tuning the communication block sizes, significantly decreasing L2 cache usage and minimizing interference from other streaming multiprocessors (SMs).

Nonetheless, assertions that utilizing PTX programming language signifies a breach of NVIDIA's CUDA dominance are flawedIndustry analysts clarify that PTX is, in fact, a component of CUDA and does not evade itThey explain that CUDA is a comprehensive software suite encompassing higher-level programming languages, an extensive collection of API libraries, and compilation tools designed for GPU programmingPTX functions as an intermediate assembly language within CUDA, closer to hardware and not typically aimed at developersConsequently, developing within CUDA generally limits control over the GPU's finer aspectsHowever, employing the PTX language allows for greater flexibility in managing underlying hardware and optimizing performance, which might explain why DeepSeek requires less computational power.

Although the DeepSeek model remains trained on NVIDIA GPUs, its demonstrated efficiency in utilizing computational resources, combined with the resulting trend of adapting domestic chips, signals significant benefit for the chip industry.

Prior to this, domestic large model companies utilized domestic chips for certain model inference or testing training tasks, albeit on a limited scale

However, spurred by the DeepSeek phenomenon, the utilization rate of domestic chips is set to rise dramatically.

Is the year for practical AI applications truly upon us? The waves generated in the upper and middle sectors will inevitably cascade downwards, and with the proliferation of the DeepSeek trend, numerous industries, including smart hardware, automotive, and finance, have begun eagerly integrating the DeepSeek model into their servicesIn the past week, the Reading Group announced that its writer assistance product, "Writer Assistant," has integrated the DeepSeek-R1 model, marking DeepSeek's initial foray into the online literature spaceThey noted that when utilizing the intelligent Q&A feature designed to assist writers with research and inspiration, DeepSeek showcases a superior understanding of intent, capable of interpreting subtext and implied meanings.

Furthermore, the extended reasoning capabilities exhibited by the R1 model serve as a noteworthy source of inspiration for online literature authors. "Online literature writers, particularly seasoned ones, frequently express dissatisfaction with the cliched and repetitive outputs of AI contentWhat they require are insights and a means to organize their thoughts," Reading Group statedFollowing the integration of DeepSeek, when authors request AI-generated outlines incorporating trending elements from a particular website, the AI not only provides a generated answer but first identifies and clearly lists relevant elements, even offering popular titles in that genre to assist writers in accessing the necessary specialized content.

Amid the competitive pressure induced by DeepSeek, OpenAI recently announced the public release of its latest o3-mini series, which also features reasoning chainsHowever, their researchers noted that while these reasoning summaries closely approach the original chains, they do not replicate them exactlyAnalysts have suggested that OpenAI's decision might stem from considerations regarding user experience, privacy protection, output quality, technical costs, and trade secrets, aiming to offer valuable insights without adverse consequences.

Back in May of the previous year, DeepSeek's low pricing sparked a "price war" among domestic AI large model providers

Advertisements

Advertisements


Leave a Reply