Voluntary Code of Conduct on the Responsible Development and Management of Advanced Generative AI Systems

September 2023

Advanced AI systems capable of generating content — such as ChatGPT, DALL·E 2, and Midjourney — have captured the world's attention. The general-purpose capabilities of these advanced AI systems offer enormous potential for innovation in a number of fields, and they are already being adopted and put to use in a variety of contexts. These advanced systems may be used to perform many different kinds of tasks — such as writing emails, answering complex questions, generating realistic images or videos, or writing software code.

While they have many benefits, advanced generative AI systems also carry a distinctly broad risk profile, due to the broad scope of data on which they are trained, their wide range of potential uses, and the scale of their deployment. Systems that are made publicly available for a range of different uses can present risks to health and safety, can propagate bias, and carry the potential for broader societal impacts, particularly when used by malicious actors. For example, the capability to generate realistic images and video, or to impersonate the voices of real people, can enable deception at a scale that can damage important institutions, including democratic and criminal justice systems. These systems may also have important implications for individual privacy rights, as highlighted in the G7 Data Protection and Privacy Authorities' Statement on Generative AI.

Generative systems can also be adapted by organizations for specific uses – such as corporate knowledge management applications or customer service tools – which generally present a narrower range of risks. Even so, there are a number of steps that need to be taken to ensure that risks are appropriately identified and mitigated.

To address and mitigate these risks, signatories to this code commit to adopting the identified measures. The code identifies measures that should be applied in advance of binding regulation pursuant to the Artificial Intelligence and Data Act by all firms developingFootnote 1 or managing the operationsFootnote 2 of a generative AI system with general-purpose capabilities, as well as additional measures that should be taken by firms developing or managing the operations of these systems that are made widely available for use, and which are therefore subject to a wider range of potentially harmful or inappropriate use. Firms developing and managing the operations of these systems both have important and complementary roles. Developers and managers need to share relevant information to ensure that adverse impacts can be addressed by the appropriate firm.

While the framework outlined here is specific to advanced generative AI systems, many of the measures are broadly applicable to a range of high-impact AI systems and can be readily adapted by firms working across Canada's AI ecosystem. It is also important to note that this code does not in any way change existing legal obligations that firms may have – for example, under the Personal Information Protection and Electronic Documents Act.

In undertaking this voluntary commitment, developers and managers of advanced generative systems commit to working to achieve the following outcomes:

  • Accountability – Firms understand their role with regard to the systems they develop or manage, put in place appropriate risk management systems, and share information with other firms as needed to avoid gaps.
  • Safety – Systems are subject to risk assessments, and mitigations needed to ensure safe operation are put in place prior to deployment.
  • Fairness and Equity – Potential impacts with regard to fairness and equity are assessed and addressed at different phases of development and deployment of the systems.
  • Transparency – Sufficient information is published to allow consumers to make informed decisions and for experts to evaluate whether risks have been adequately addressed.
  • Human Oversight and Monitoring – System use is monitored after deployment, and updates are implemented as needed to address any risks that materialize.
  • Validity and Robustness – Systems operate as intended, are secure against cyber attacks, and their behaviour in response to the range of tasks or situations to which they are likely to be exposed is understood.

Signatories also commit to support the ongoing development of a robust, responsible AI ecosystem in Canada. This includes contributing to the development and application of standards, sharing information and best practices with other members of the AI ecosystem, collaborating with researchers working to advance responsible AI, and collaborating with other actors, including governments, to support public awareness and education on AI. Signatories also commit to develop and deploy AI systems in a manner that will drive inclusive and sustainable growth in Canada, including by prioritizing human rights, accessibility and environmental sustainability, and to harness the potential of AI to address the most pressing global challenges of our time.

Frequently asked questions for the Voluntary Code of Conduct on Advanced Generative AI Systems.

Signatories to the code of conduct

Signatories
Organisme d'autoréglementation du courtage immobilier du Québec (OACIQ)
CGI
IBM
Scale AI
Protexxa Inc.
BlueDot
Resemble AI
Ada
AlayaCare
Alberta Machine Intelligence Institute (Amii)
AltaML
Appen
BlackBerry
Council of Canadian Innovators
Cohere
Coveo
kama.ai
Mila
OpenText
Responsible Artificial Intelligence Institute
Ranovus
TELUS
Vector Institute

Measures to be undertaken pursuant to the Code of Conduct

Principle Measures All advanced generative systems Advanced generative systems available for public use
Developers Managers Developers Managers
Accountability Implement a comprehensive risk management framework proportionate to the nature and risk profile of activities. This includes establishing policies, procedures, and training to ensure that staff are familiar with their duties and the organization's risk management practices. Yes Yes Yes Yes
Share information and best practices on risk management with firms playing complementary roles in the ecosystem. Yes Yes Yes Yes
Employ multiple lines of defence, including conducting third-party audits prior to release. No No Yes No
Safety Perform a comprehensive assessment of reasonably foreseeable potential adverse impacts, including risks associated with inappropriate or malicious use of the system. Yes Yes Yes Yes
Implement proportionate measures to mitigate risks of harm, such as by creating safeguards against malicious use. Yes No Yes No
Make available to downstream developers and managers guidance on appropriate system usage, including information on measures taken to address risks. Yes No Yes No
Fairness and Equity Assess and curate datasets used for training to manage data quality and potential biases. Yes No Yes No
Implement diverse testing methods and measures to assess and mitigate risk of biased output prior to release. Yes No Yes No
Transparency Publish information on capabilities and limitations of the system. No No Yes No
Develop and implement a reliable and freely available method to detect content generated by the system, with a near-term focus on audio-visual content (e.g., watermarking). No No Yes No
Publish a description of the types of training data used to develop the system, as well as measures taken to identify and mitigate risks. No No Yes No
Ensure that systems that could be mistaken for humans are clearly and prominently identified as AI systems. No Yes No Yes
Human Oversight and Monitoring Monitor the operation of the system for harmful uses or impacts after it is made available, including through the use of third-party feedback channels, and inform the developer and/or implement usage controls as needed to mitigate harm. No Yes No Yes
Maintain a database of reported incidents after deployment, and provide updates as needed to ensure effective mitigation measures. Yes No Yes No
Validity and Robustness Use a wide variety of testing methods across a spectrum of tasks and contexts prior to deployment to measure performance and ensure robustness. Yes No Yes No
Employ adversarial testing (i.e., red-teaming) to identify vulnerabilities. Yes No Yes No
Perform an assessment of cyber-security risk and implement proportionate measures to mitigate risks, including with regard to data poisoning. Yes No Yes Yes
Perform benchmarking to measure the model's performance against recognized standards. Yes No Yes No