ARTICLE
26 March 2019

Lowering The "Creep" Factor: Canada's New Directive On Automated Decision-Making

sL
Lawson Lundell LLP

Contributor

Lawson Lundell is a leading full-service law firm, known for our strategic approach to legal services. With over 160 lawyers, and offices in Vancouver, Calgary, Yellowknife and Kelowna, we are widely recognized for our depth of experience and innovative solutions to complex business law and litigation matters across various sectors.
Organizations are implementing Artificial Intelligence (AI) in numerous ways, directly and indirectly, to save time and money.
Canada Privacy

Organizations are implementing Artificial Intelligence (AI) in numerous ways, directly and indirectly, to save time and money.

How can an organization ensure the AI platform or service fairly and adequately meets its needs, without being too privacy intrusive, biased, and uncertain?

A new Canadian method for managing these issues is articulated in a new Federal Directive on Automated Decision Making, the AIA: Algorithmic Impact Assessment. The AIA has similarities to the PIAs (Privacy Impact Assessments) privacy professionals have been conducting for years.

This Directive is meant as a guide for the Federal Government's use of AI in an ethical, responsible and transparent manner. It is to assist in the determination of whether a particular AI can contribute to government administrative decisions.

The Directive addresses "Automated Decision Systems", which include any information technology designed to either assist or replace the judgment of a human decision maker. The Directive provides for an AIA to be completed prior to the production of any Automated Decision System to be used in federal administration. The AIA itself is an interactive questionnaire to help organizations understand and mitigate risks, including what kind of human intervention and monitoring an AI tool will require.

While the Directive only applies to federal government institutions and AI vendors for federal government institutions, other governmental and private organizations can look to this model as a standard for transparency in AI procurement. We anticipate the AIA will be adopted by other governments, establishing a common Canadian standard.

An AIA process will enable organizations to better establish trust with their clients and workforce. AIAs will help organizations explain and defend their decisions, if necessary. It is not always clear how a decision is rendered through an algorithm. If the decision is based on historical information or "big data" it will not be clear whether that is biased. These raise privacy and human rights issues which give rise to various risks to the organizations using the AI. Like a PIA, the AIA will help identify the level of risk and what sort of mitigation is appropriate.

While few are asking what is behind AI today, those questions are coming.

The content of this article is intended to provide a general guide to the subject matter. Specialist advice should be sought about your specific circumstances.

Mondaq uses cookies on this website. By using our website you agree to our use of cookies as set out in our Privacy Policy.

Learn More