The Online Streaming Act (Bill C-11) claims to level the playing field … but for whom? Nawshaba Ahmed and I ask, in our piece published in The Conversation.
See also my comments to the Heritage committee on algorithmic recommendations and C-11.
The Online Streaming Act (Bill C-11) claims to level the playing field … but for whom? Nawshaba Ahmed and I ask, in our piece published in The Conversation.
See also my comments to the Heritage committee on algorithmic recommendations and C-11.
44th Parliament, 1st Session
Tuesday, May 31, 2022, 6:30 p.m. to 8:30 p.m.
Watch on ParlVu
My name is Sara Bannerman (she/her). I’m a Canada Research Chair in Communication Policy and Governance at McMaster University.
Today I will focus on discoverability and algorithmic bias.
Governments around the are working on measures to ensure that algorithms are accountable.
There is a common misconception that streaming platforms’ recommend what the user wants to see.
Platforms show us what they want to show us.
They show us what will keep us watching ads, purchasing advertised products, contributing
data, and subscribing.
Platforms are not neutral. They serve
their business interests. As well, there are three types of bias I am concerned
about. These biases can affect both users and content providers.
First, there could be bias if algorithms are used to select content for carriage on a streaming service--
The research in this area is only recently developing.
CRTC intervention in the algorithms raises many difficult problems. It may not be the first, most likely, or best answer to them. The CRTC has said, today, it does not want to play that role. But the Commission could play a role in bringing such problems to light.
There are concerns that requiring discoverability would
infringe on freedom of expression.
Streaming services’ user interfaces and recommendations may be forms of
expression, and, if so, regulatory interventions could constitute a limit on
that expression.
There are legitimate concerns that promoting some content could mean demoting other
content, among other concerns.
Sometimes limits on expression are justified, but they must be justified.
To understand whether any justification exists, or even just to understand, we
need data.
It may be that the best role for the CRTC will be to monitor
and call attention to problems, not just with the discoverability of Canadian
content, but also with recommender biases relating to other Canadian values, so
that civil society and others can intervene.
The CRTC can only do that if it has data.
The provisions on information-provision and information disclosure in the bill are important to the study and examination of discoverability algorithms and data, and to the CRTC’s potential work with outside organizations on this matter.
It may be necessary to require platforms to collect certain data to permit these examinations to happen.The Canadian broadcasting system has often served dominant
groups.
It has also been open to change and improvement based the work of civil society
and others.
We need to ensure that the discoverability mechanisms of online streaming platforms are also open to critique and change through public transparency, debate, and data.
Thank you.
9 (1) (n) the collection and provision to the Commission […]
9 (1) (o) the collection and provision to the Commission […]
9 (1) (o) (v) information related to discoverability; and [NEW]
I object to Netflix’s proposed amendments to Bill C-10 section 25.3, which could equally be proposed to be applied to C-11.
If their proposed 25.3 (4 and 5) (a) (a.1) is adopted (I do not support this), it should read “must not knowingly be publicly disclosed or made available […].” This would permit the CRTC working privately with groups to examine discoverability mechanisms and other matters.