The AGI Race Between the US and China Doesn’t Exist.
When I write “China”, I refer to the political and economic entity, the People’s Republic of China, founded in 1949.
Leading US AI companies are currently rushing towards developing artificial general intelligence (AGI) by building and training increasingly powerful large language models (LLMs), as well as building architecture on top of them. This is frequently framed in terms of a great power competition for technological supremacy between the US and China.
However, China has neither the resources nor any interest in competing with the US in developing artificial general intelligence (AGI) primarily via scaling Large Language Models (LLMs).
In brief, China does not compete with the US in developing AGI via LLMs because of:
Resources and technology: China does not have access to the computational resources1 (compute, here specifically data centre-grade GPUs) needed for large-scale training runs of large language models. This gap will only widen over time; China is failing to develop a domestic semiconductor industry, despite massive efforts to do so, and is increasingly cut off from international semiconductor supply chains.
Political goals: The Chinese Communist Party (CCP) is above all concerned with ensuring social stability and staying in power. It will not tolerate large language models that cannot reliably be controlled and prevented from leaking sensitive information, raising politically sensitive topics, or challenging the regime. Reliably prohibiting LLMs from doing any of these things is an unsolved technical problem. This has and will continue to stifle the development and deployment of Chinese LLMs for the Chinese market, as a single political faux pas will spell doom for the companies involved.
Therefore, as it currently stands, there is no AGI race between China and the US.
China Cannot Compete with the US to Reach AGI First Via Scaling LLMs
Today, training powerful LLMs like GPT-4 requires massive GPU clusters. And training LLMs on a larger number of GPUs is the most common and reliable way of increasing their capabilities. This is known as the bitter lesson in machine learning; increasing the compute a model is trained on leads to higher progress in the model's capabilities than any changes in the algorithm's architecture. So if China wants to “win the AGI race” against the US, it would need to maximise its domestic GPU capacities. However, China lacks domestic semiconductor production facilities for producing data centre-grade GPUs. So to increase domestic GPU capacities, it has to import GPUs or semiconductor manufacturing equipment (SME).
However, since the US introduced its export controls on advanced chips and SMEs in October 2022, China can no longer purchase GPUs or SMEs abroad. And with the semiconductor industry advancing at lightning speed, China is falling further and further behind every day2. These policies are just the most recent ones in a row of ongoing measures by the US restricting the transfer of semiconductor technology to Chinese companies, beginning with Huawei in 2019. In early 2023, the Netherlands and Japan, which play a key role in the global semiconductor manufacturing chain, joined the US and introduced SME export controls targeting China. Crucially, these combined export controls apply to ASML’s most advanced lithography machines, the only machines in the world which can produce cutting-edge chips.
The problem for China lies not so much in designing cutting-edge GPUs, but in designing and building SMEs to produce the advanced GPUs needed for training LLMs, a far more challenging task. But China has been trying - and failing - to build domestic semiconductor manufacturing chains for cutting-edge chips since at least the 1990s. In the 1990s, the state-sponsored Projects 908 and 909 only resulted in one company that today produces chips using 55nm process technology (technology from the mid-2000s, producing less complex chips used for applications like the Internet of Things and electric vehicles) and occupies around 3% of global foundry market share. In its 13th Five-Year Plan (2016-2020), the CCP listed building a domestic semiconductor industry as a priority, and in the 14th Five-Year Plan (2021-2025), developing domestic semiconductor supply chains appears as one of the country’s main goals in industrial and economic policy. Since then, China has increased domestic chip production capacities, but still only succeeds at producing larger, less complex chips. And achieving even this success was possible largely due to the forced transfer of intellectual property from US-American and Taiwanese semiconductor firms. This is now much harder with strict export controls from the US on advanced GPUs and a broader plethora of measures aimed at restricting the transfer of chips and SME technology and design to Chinese firms.
In addition, in the summer of 2022, the Chinese government started investigating multiple senior figures who were responsible for industrial policy and state investment programmes to create a domestic cutting-edge chip industry starting in 2014. These investigations are based on “opaque allegations” of officials breaking the law. In early 2023, China then changed its strategy from granting generous subsidies to semiconductor manufacturing companies to focusing on a small number of potential national champions. This recent policy shift indicates that the Chinese leadership has decided that prior strategies have failed. In addition, the new policy programme has a smaller budget, which points to the Chinese government prioritising other challenges (more on this below).
On a general note, Chinese industrial policy officially has the goal of building industry and increasing economic growth, but its actual goal is first and foremost to preserve CCP power and social stability. Chinese industrial policy routinely fails at its stated goals precisely because it is optimised for preserving CCP power, which it has succeeded at since the late 1970s. This goes beyond the semiconductor industry and e.g. also causes failures in China’s green transition. For these reasons, the probability that the Chinese government will succeed at or even properly try to build a domestic semiconductor industry capable of producing cutting-edge, data centre-grade GPUs is extremely low.
On top of everything, China is facing food security, public health, and demographic challenges, and the rising cost of Chinese labour is threatening its economic model as the world's workbench. These combined challenges pose a more significant threat to the CCP’s hold on power than lagging behind the USA in AGI, so the CCP will most likely prioritise these challenges and continue to reduce funds available for building domestic GPU supply chains. And the challenges China faces today will only get more serious, as its population ages further and climate change increases the frequency of extreme weather events. Hence, in combination with the growing lead of the US in semiconductor technology, as time passes, it will become even more difficult for China to catch up to the US than it is today.
For Political Reasons, China Won’t Compete with the US to Reach AGI First Via Scaling LLMs
The CCP is not interested in reaching AGI by scaling LLMs. Reliably controlling powerful LLMs is an unsolved technical problem3: for instance, it is currently impossible to get an LLM to reliably never output specific words or phrases or touch on certain topics. Given this, any sufficiently powerful Chinese LLM is guaranteed to make a politically taboo statement sooner or later, like referencing the Tian’anmen Square massacre. This makes this line of AI development too politically sensitive and ultimately undesirable in the Chinese environment.
Instead, the CCP wants and needs reliable, narrow, controllable AI systems to censor internet discourse, improve China’s social credit system, and expand citizen surveillance. This is reflected in the topics Chinese and US AI researchers write about. Chinese researchers are overrepresented among highly cited papers on AI applications in surveillance and computer vision, while US researchers publish more leading papers on deep learning.
Considering this, one of two things will happen: Either Chinese LLMs will continue to be weaker, due to a lack of compute or by design to comply with government guidelines, or a Chinese developer will release a powerful LLM on the Chinese internet, it will cause a political scandal by not conforming with censorship rules or state-issued guidelines, and the government will shut it down.
Chinese companies will only build LLMs that reliably comply with Chinese censorship laws and upcoming laws or guidelines on generative AI. To ensure models never go off-script, companies will have to make them weaker, and very carefully control who gains access to the models or the APIs. For example, the March 2023 demo for Baidu’s new generative language model Ernie Bot consisted only of pre-recorded video clips, disappointing the public, and so far, only businesses can apply for licenses to use the model. Baidu advertised Ernie Bot as a multimodal model, meaning it should be able to generate responses to different forms of input, e.g. including text and images. However, Ernie Bot lacks the ability to interpret images and generate a text response. Furthermore, users who had the opportunity to test Ernie Bot report that it performed slightly worse than GPT-3, a model that is now almost three years old, and, unlike ChatGPT, shirks away from answering questions on political topics. For context, these reviews were published in late March 2023, when OpenAI had already released the more powerful GPT-4. One reviewer summarised the situation by commenting on the model’s “bearable mediocrity” and writing it is “probably good enough for the Chinese market”. In short, Ernie Bot is just good enough to probably for now satisfy the demand in a market where all serious competitors are banned. In direct competition with the generative language models from companies like OpenAI, Ernie Bot would not be able to compete.
If a Chinese developer ever releases an LLM that is so powerful it inevitably oversteps censorship rules at some point, the Chinese government will block it and crack down on the company that released it. In the past, the CCP has held tech companies accountable for any undesirable content posted by users of their platforms, and according to the new draft guidelines on generative AI, it will do the same regarding LLMs. The temporary disappearance of Jack Ma in 2020 when the CCP decided that his company Alibaba had become too powerful is another cautionary tale for Chinese tech CEOs to not challenge the CCP.
Indeed, the new drafts on guidelines for generative AI in China from April 20234, published shortly after I first wrote this piece, give an idea of the intentions of the CCP regarding the regulation of generative AI models, including LLMs. Specifically, these draft guidelines state that “the content generated by generative artificial intelligence [...] must not contain subversion of state power”. This statement is a clear indicator of the CCP’s intention to prohibit any generative AI model from being deployed that may at some point produce politically undesirable outputs. But So under the draft guidelines, powerful LLMs would be completely prohibited.
No Global AGI Race in Sight
In conclusion, China is not in a position to build AGI via scaling LLMs because it lacks the resources, technology, and political will to do so.
First, China doesn’t have the necessary computing resources, data centre-grade GPUs, to continue doing increasingly large LLM training runs. China is unable to purchase these GPUs, primarily designed in the US and manufactured outside of China, because the US and its allies have placed strict export controls on this technology, as well as the SMEs needed to produce sufficiently powerful GPUs.
Second, China is also unable to manufacture data centre-grade GPUs domestically. This situation is only worsening as the CCP will have to deprioritise industrial policy programmes to develop domestic cutting-edge SMEs and GPUs in favour of dealing with more pressing challenges, such as demographic change and the rising cost of Chinese labour.
Lastly, the Chinese government has no interest in reaching AGI via powerful LLMs, since LLMs cannot be controlled reliably. The Chinese government made this clear in its recent draft guidelines on generative AI. Instead, the CCP is interested in using narrow, controllable AI systems for surveillance. As a result, AI companies in China will either continue to release weak LLMs, like Baidu’s Ernie Bot or will be cracked down on if they release a sufficiently powerful and therefore uncontrollable model.
In short, there is no AGI race between the USA and China. Hence, framing the development of AGI as a competition between the USA and China, in which the US must compete to win, is misleading. All companies with leading AGI ambitions are located in the US and do not have direct competitors in China. If concerns over a US-China AGI race are the reason why your company is accelerating its progress towards AGI, then your concerns are unfounded and you should stop.
In March 2023, Nvidia introduced its new GPU unit for the Chinese market, called the H800. Compared to its most recent product for the Western market, the A800, the H800 performs the same number of operations per second but has a much lower interconnect bandwidth, lower even than Nvidia’s last-generation GPU unit, the A100, used to train models like GPT-3. With these features, the H800 will perform well for AI models similar in size to GPT-3, but the larger the model, the higher the GPU’s interconnect bandwidth needed to train it efficiently. Hence, the export controls are not having a very strong impact on China’s ability to train large language models right now, as Chinese companies can still purchase the H800, but will keep China from training models significantly larger than the ones that exist today, effectively putting a hard limit on Chinese innovation potential. For illustration, see this diagramme by Lennart Heim.
There is currently no solution to jailbreak prompts, adversarial attacks, or other similar methods to surpass filtering techniques of unwanted content in LLMs.