BACKGROUND

The Ministry of Electronics and Information Technology (MeitY) reportedly issued an advisory on 1 March 2024 (Advisory) to "intermediaries" and "platforms" hosting artificial Intelligence (AI) including generative AI based models.

The Information Technology Act, 2000 (IT Act) defines the term "intermediary", which in simple terms refers to an entity that on behalf of another person deals with or provides a service in relation to an electronic record and typically includes players like telecom service providers, cloud-hosting providers, e-commerce platforms, etc. However, as the term '"platform" is not explicitly defined under the IT Act or Information Technology (Intermediary Guidelines and Digital Media Ethics Code) Rules, 2021 (Intermediary Guidelines) , the extent to which the Advisory applies to digital "platforms" deploying AI based models still remains ambiguous.

The Central Government (Government) has, during press interactions following the issuance of the Advisory, clarified that the Advisory does not aim to govern startups but only "significant platforms", but a formal clarification or amendment to this effect is yet to be released.

KEY HIGHLIGHTS OF THE ADVISORY

Restriction on use for prohibited information / content:

"Intermediaries" and AI based "platforms" are obligated to ensure that any AI tool on its "platform" or being used through its computer resource, should not deal with unlawful or wrongful content that threatens the integrity / security of India or its relations with other countries, pornographic material, false or misleading information, infringing material, etc. With the upcoming general elections in India, it is likely that the Government is aiming to curb instances of fake news and misleading information circulated through digital platforms in the country, particularly in the period leading up to the general elections.

Steps to prevent bias in the integrity of electoral process:

"Intermediaries" and AI based "platforms" are required to ensure that their "platform" does not permit any bias / discrimination / threat to the integrity of the electoral process through the use of AI tools. This means appropriate mechanisms may have to be implemented by "intermediaries" to promptly detect and address any such instances, especially when the content is generated through AI tools.

Explicit approval of the Government before using and making available under-testing/ unreliable AI Tools in India:

The Advisory mandates that the "intermediaries" and AI "platforms" take explicit permission of the Government prior to using and making available AI Tools that are un-tested and unreliable. Notably, the foregoing terms are not defined under the IT Act and the Intermediary Guidelines and hence it will have to be seen what interpretation is accorded to them. As previously noted, Government has clarified that the Advisory is aimed at governing "significant platforms" and does not govern startups. However, in absence of a clear-cut definition of "significant platform" under the law, a formal clarification is awaited in this regard. Additionally, the Advisory requires the "intermediaries" and AI "platforms" to affix appropriate labels regarding the possible and inherent fallibility or unreliability of the output generated through such AI Tools, with an aim to clarify to the users that information may not be reliable and correct in all cases. The Advisory suggests that this can be reinforced by the use of "consent pop-ups" to inform the users.

Clear communication to the users about the consequences of dealing with unlawful information:

The Advisory requires the "intermediaries" and digital "platforms" deploying AI to communicate to all of their users, through the terms of services and user agreements, about the consequence of dealing with the unlawful information on its "platform" which may include disabling of access to or removal of non-compliant information, termination or suspension of access to the user account, and punishment under applicable law.

Appropriate labelling of information that may be used as misinformation or deepfake:

The Advisory provides that any intermediary that permits or facilitates synthetic creation / generation / modification of any information (be it text, audio, visual or audio-visual), which can potentially be used as "misinformation" or "deepfake", should label such information or embed it with a permanent unique metadata or identifier. The purpose of such labelling would be to (a) identify that such information has been modified in the course of dissemination and (b) identify the creator / first originator of such "misinformation" or "deepfake".

Filing of action taken report:

The Advisory requires all "intermediaries" to ensure compliance with the steps above in the form of an "action taken-cum-status" report within 15 (fifteen) days of the Advisory.

Comment

The Advisory reflects the Government's intention to regulate the rapidly growing field of AI and ensure its ethical utilisation which is the need of the hour, especially in the absence of any dedicated legal framework at present. It also hints towards the Government's inclination towards mitigating growing concerns surrounding data privacy, security, and implications of AI algorithms. While there is uncertainty and ambiguity around the legal basis and applicability of the Advisory, there is no doubt that the Advisory has caused ripples in the industry. It would be interesting to see how the Government works towards providing more clarity on the testing threshold for the AI based models and defining the standards that the "intermediaries" / AI "platforms" are required to maintain while providing services to their users in India. It will also have to be seen how these aspects are addressed under the proposed Digital India Act, which is expected to be rolled out in the latter part of this year.

The content of this document do not necessarily reflect the views/position of Khaitan & Co but remain solely those of the author(s). For any further queries or follow up please contact Khaitan & Co at legalalerts@khaitanco.com