Explore the new Chinese video-generating AI model that censors politically sensitive topics deemed controversial by the Chinese government
Kuaishou, a Beijing-based company, introduced the Kling model to waitlist access earlier this year for consumers with Chinese phone numbers. It was implemented today for individuals willing to submit their email addresses. Users can input prompts after registering to have the model produce five-second videos of the content they have described.
Kling functions substantially as advertised. Its 720p videos, which require approximately one minute to produce, adhere closely to the instructions. Kling appears to simulate physics, such as the rustling of foliage and water flow, to a similar extent as video-generating models such as Gen-3 from AI startup Runway and Sora from OpenAI.
However, Kling explicitly refuses to produce recordings regarding specific topics. A nonspecific error message is generated in response to prompts such as “Democracy in China,” “Chinese President Xi Jinping walking down the street,” and “Tiananmen Square protests.”
It seems that the filtering is occurring exclusively at the prompt level. Kling can animate still images and generate a video of a portrait of Jinping without complaint, provided that the accompanying prompt does not reference Jinping by name (e.g., “This man is giving a speech”).
Kuaishou has been contacted for comment.
Kling’s curious behavior is likely due to the Chinese government’s heavy political pressure on generative AI initiatives in the region.
The Financial Times reported earlier this month that the Cyberspace Administration of China (CAC), China’s foremost internet regulator, will evaluate AI models in China to guarantee that their responses to sensitive topics “embody core socialist values.” According to the Financial Times report, CAC officials are to evaluate models based on their responses to diverse inquiries, including numerous inquiries regarding Jinping and criticism of the Communist Party.
According to reports, the CAC has even gone so far as to suggest a blacklist of sources that are ineligible for training AI models. Companies that submit models for evaluation must develop tens of thousands of questions intended to evaluate the models’ ability to generate “safe” responses.
The outcome is AI systems that refrain from responding to inquiries that could provoke Chinese regulators’ disapproval. The BBC discovered last year that Ernie, the flagship AI chatbot model of Chinese company Baidu, demurred and deflected when posed questions that could be perceived as politically controversial, such as “Is Xinjiang a good place?” or “Is Tibet a good place?”
China’s AI advancements are at risk of being impeded by the draconian policies. In addition to necessitating the sifting of data to eliminate politically sensitive information, they also necessitate allocating a significant amount of development time to establishing ideological guardrails, which may still fail, as Kling demonstrates.
From the user’s perspective, China’s AI regulations are already resulting in two distinct classes of models: severely restricted by rigorous filtering and significantly less so. Is this truly beneficial for the broader AI ecosystem?
After six months of bearish stagnation, the Worldcoin (WLD) price is set to rise following the global adoption of the…
The crypto confidentiality network COTI introduces new Layer-2 testnet months after its Devnet launch to start the third phase of…
The new app Nurture is made for kids ages 4 to 7 and has games and interactive material to keep…
Bitget, a cryptocurrency exchange, and Foresight Ventures, an investment firm specializing in Web3 technology, have invested $30 million in The…
The crypto space currently needs more details regarding the services that World Liberty Financial will provide. However, this could result…
Five entities and three individuals have been accused by the US Securities and Exchange Commission (SEC) of the operation of…