Friday, 19 July 2024

Crypto News

China sets stricter rules for training generative AI models

China sets stricter rules for training generative AI models

China has released draft security regulations for companies providing generative artificial intelligence (AI) services, encompassing restrictions on data sources used for AI model training.

On Wednesday, Oct. 11, the proposed regulations were released by the National Information Security Standardization Committee, comprising representatives from the Cyberspace Administration of China (CAC), the Ministry of Industry and Information Technology and law enforcement agencies.

Generative AI, as exemplified by the accomplishments of OpenAI’s ChatGPT chatbot, acquires the ability to perform tasks through the analysis of historical data and generates fresh content such as text and images based on this training.

Screenshot of the National Information Security Standardization Committee (NISSC) publication. Source: NISSC

The committee recommends performing a security evaluation on the content utilized to train publicly accessible generative AI models. Content exceeding “5% in the form of unlawful and detrimental information” will be designated for blacklisting. This category includes content advocating terrorism, violence, subversion of the socialist system, harm to the country’s reputation and actions undermining national cohesion and societal stability.

The draft regulations also emphasize that data subject to censorship on the Chinese internet should not serve as training material for these models. This development follows slightly more than a month after regulatory authorities granted permission to various Chinese tech companies, including the prominent search engine firm Baidu, to introduce their generative AI-driven chatbots to the general public.

Since April, the CAC has consistently communicated its requirement for companies to provide security evaluations to regulatory bodies before introducing generative AI-powered services to the public. In July, the cyberspace regulator released a set of guidelines governing these services, which industry analysts noted were considerably less burdensome compared to the measures proposed in the initial April draft.

Related: Biden considers tightening AI chip controls to China via third parties

The recently unveiled draft security stipulations, necessitate that organizations engaged in training these AI models obtain explicit consent from individuals whose personal data, encompassing biometric information, is employed for training. Additionally, the guidelines include comprehensive instructions on preventing infringements…

Click Here to Read the Full Original Article at News…