*Insufficiently Frequently Asked Questions
Eager Student has some questions for Scholarly Lawyer about
the UK government’s Online Harms White Paper.
ES. Why are people so perturbed about the proposed Online
Harms legislation?
SL. Most people aren’t, even if they have heard of it. But
they should be.
ES. What can be wrong with preventing harm?
SL. Indeed, who could be against that? But this is a
proposal for legislation. We have to peel off the label and check what is in
the bottle. When we look inside, we find
a wormhole back to a darker age.
ES. A darker age?
SL. Consider the days when unregulated theatres were reckoned
to be a danger to society and the Lord Chamberlain censored plays. That power
was abolished in 1968, to great rejoicing. The theatres were liberated. They
could be as rude and controversial as they liked, short of provoking a breach
of the peace.
The White Paper proposes a Lord Chamberlain for the internet.
Granted, it would be an independent regulator, similar to Ofcom, not a royal
official. It might even be Ofcom itself. But the essence is the same. And this time the target would not be a
handful of playwrights out to shock and offend, but all of us who use the
internet.
ES. This is melodramatic, surely. How could a digital
Lord Chamberlain censor the tens of millions of us who use the internet and
social media?
SL. Would the regulator personally strike a blue pencil through
specific items of content? No. But the legislation is no less censorious for
that. It would co-opt internet platforms and search engines to do the work,
under the banner of a duty of care articulated and enforced by the regulator.
ES. Doesn’t the internet route around censorship?
SL. Once your post or tweet has passed through your ISP’s
system it is on the internet, chopped up into fragments and bouncing around an
interconnected mesh of routers: very difficult to isolate and capture or block.
But your own ISP has always been a choke point. It can be
turned into a gatekeeper instead of a gateway. That danger was recognised early
on when Article 15 of the ECommerce Directive prevented EU governments from doing
that.
Nowadays we communicate via large platforms: social media
companies, messaging apps and the rest. Our tweets, posts and instant messages
end up in a few databases. So we have new choke points. Companies may be able to block, filter and
take down users’ content on their platforms, even if broad brush rather than
blue pencil. Our digital Lord Chamberlain would commandeer that capability at
one remove,
ES. Even so, where is the censorship? This is about online
safety, not suppressing controversy.
SL. The particular vice at the heart of the White Paper is
the latitude for the regulator to deem things to be harmful. If the proposals
were only about safety properly so-called, such as risk of death and personal
injury, that would correspond to offline duties of care and draw a clear line. Or
if the proposals were only about unlawful online behaviour, the built-in line
between lawful and unlawful would provide some protection against overreach.
But the proposals are neither, and they do not.
ES. If something is not presently unlawful, would the
White Paper proposals make it so?
SL. Yes and no. Asserting that you have been harmed by
reading something that I have posted online is not a basis for a legal claim against me, even if you are shocked, outraged, offended or distressed by it. There
would have to be more, such as incitement or harassment. That is the same rule
as for publishing offline.
But the White Paper’s duty of care could require a platform
to take action against that same post, regardless of its lawfulness. It might
extend to blocking, filtering or removing. The platform could be fined if it
didn’t have good enough systems for reducing the risk of harm that the
regulator deemed to exist from that kind of content.
ES. So the platform could face legal consequences for
failing to censor my material, even though I have done nothing unlawful by
writing and posting it?
SL. Exactly so.
ES. I’m curious as to why would anyone would set up a new
regime to address behaviour that is already unlawful.
SL. To create a new law enforcement route. A digital Lord
Chamberlain would co-opt online intermediaries to the task: mixing our
metaphors horribly, an online sheriff riding out with a posse of conscripted deputies
to bring order to Digital Dodge. Whether that is a good way of going about
things is a matter of highly contested opinion.
But the genesis of the White Paper was in something quite
different from unlawful behaviour: lawful online content and behaviour said to pose
a risk of harm, whether to an individual user or to society.
ES. Is it really right that the regulator could treat almost
anything as harmful?
SL. The White Paper proposes to impose a duty of care on online
intermediaries in respect of harm. It offers no definition of harm, nor what
constitutes risk of harm. Apart from a few specific exclusions, that would be open
to interpretation by the new online regulator.
ES. A duty of care. Like health and safety?
SL. The name is similar, but the White Paper’s duty of care
bears scant resemblance to comparable offline duties of care.
ES. How does it differ?
SL. In two main ways. First, corresponding duties of care
that have been developed in the offline world are limited to safety in the acknowledged
sense of the word: risk of death, personal injury and damage to physical
property. Those are objective concepts
which contribute to keeping duties of care within ascertainable and just limits.
Universal, generalised duties of care do not exist offline.
The White Paper duty of care is not limited in that way. It
speaks only of harm, undefined. Harm is a broad concept. Harm and safety become
fluid, subjective notions when applied to speech. We can find ourselves asserting
that speech is violence and that we are entitled to be kept safe not just from
threats and intimidation, but from opinions and facts that we find disagreeable.
The online regulator could adopt the standard of the most
easily offended reader. It could re-invent blasphemy under the umbrella of the
duty of care. It could treat the internet like a daytime TV show. Ofcom, let us
not forget, suggested to survey participants in 2018 that bad language is a “harmful
thing”. In last year’s survey it described “offensive language” as a “potential
harm”.
Undefined harm, when applied to speech, is not just in the
eye and ear of the beholder but in the opinion of the regulator. The duty, and thus the regulator’s powers,
would deliberately go far wider than anything that is contrary to the law.
ES. So even if our digital Lord Chamberlain is not censoring
specific items, it still gets to decide what kind of thing is and is not
harmful?
SL. Yes. Writing a
Code of Practice can be as censorious as wielding a blue pencil.
ES. And the second difference from an offline duty of
care?
SL. Health and safety is about injury caused by the occupier
or employer: tripping over a loose floorboard, for instance. A duty of care would not normally extend to injury
inflicted by one visitor on another. It could do if the occupier or employer
created or exacerbated the specific risk of that happening, or assumed
responsibility for the risk.
But least of all is there an offline duty of care in respect
of what one visitor says to another. That is exactly the kind of duty that the
White Paper proposals would impose on online intermediaries.
The White Paper is twice removed from any comparable offline
duty of care.
ES. Could they not define a limited version of harm?
SL. John Humphrys
suggested that on the Today programme back in April last year. The New Zealand
Harmful Digital Communications Act passed in 2015 did define harm: “serious
emotional distress”. That has the merit of focusing on a significant aspect of
the mischief that is so often associated with social media: bullying,
intimidation, abuse and the rest. But that would be a small sliver of the
landscape covered by the White Paper.
Recently one of the progenitors of the plan for a duty of
care, the Carnegie UK Institute, published its own suggested draft Bill. It does not define or limit the concept of harm.
ES. Is there a disadvantage in defining harm?
SL. That depends on your perspective. A clear, objective
boundary for harm would inevitably shrink the regulator’s remit and limit its
discretion. It would have to act within the boundaries of a set of rules. So if
your aim is to maximise the power and flexibility of the regulator, then you don’t
want to cramp its style by setting clear limits. From this viewpoint vagueness
is a virtue.
Opinions will differ on whether that is a good or a bad
thing. Traditionally the consensus has been that individual speech should be
governed by rules, not discretion. Discretion, it might be thought, is a
defining characteristic of censorship.
The very breadth of the territory covered by the White Paper
and its predecessor Internet Safety Green Paper may explain why the objection to
undefined harm is so steadfastly ignored. A catch-all is no longer a catch-all
once you spell out what it catches.
ES. You say this is about individual speech, but isn’t the
point to regulate the platforms?
SL. It is true that the intermediaries would be the direct
subject of the duty of care. They would suffer the penalties if they breached
the duty. But we users are the ones who would suffer the effects of our online
speech being deemed harmful. Any steps that the platforms were obliged to take would
be aimed at what we do and say online. Users are the ultimate, if indirect,
target.
ES. I have seen it said that the duty of care should focus
on platforms’ processes rather than on specific user content. How might that
work?
SL. Carnegie has criticised the White Paper as being overly
focused on types of content. It says the White Paper opened up the government
to “(legitimate) criticism from free speech campaigners and other groups that
this is a regime about moderation, censorship and takedown”. Instead the regime
should be about designing services that “hold in balance the rights of all
users, reduce the risk of reasonably foreseeable harms to individuals and
mitigate the cumulative impact on society.”
To this end Carnegie proposes a ‘systemic’ approach: “cross-cutting codes which focus on process and the routes to likely harm”. For the most part
its draft Bill categorises required Codes of Practice not by kinds of problematic
speech (misinformation and so on) but in terms of risk assessment, risk factors
in service design, discovery and navigation procedures, how users can protect
themselves from harm, and transparency. There would be specific requirements on
operators to carry out risk assessments, risk minimisation measures, testing
and to be transparent.
Carnegie has said: “The regulatory emphasis would be on what
is a reasonable response to risk, taken at a general level. In this, formal
risk assessments constitute part of the harm reduction cycle; the
appropriateness of responses should be measured by the regulator against this.”
ES. Would a systemic approach make a difference?
SL. The idea is that a systemic approach would focus on
process design and distance the regulator from judgements about content. The
regulator should not, say Carnegie, be taking decisions about user complaints
in individual cases.
But risk is not a self-contained concept. We have to ask:
‘risk of what?’ If the only answer provided is ‘harm’, and harm is left undefined,
we are back to someone having to decide what counts as harmful. Only then can we
measure risk. How can we do that without considering types of content? How can
the regulator measure the effectiveness of intermediaries’ harm reduction
measures without knowing what kind of content is harmful?
What Carnegie has previously said about measuring harm
reduction suggests that the regulator would indeed have to decide which types
of content are to be regarded as harmful: “…what is measured is the incidence
of artefacts that – according to the code drawn up by the regulator –
are deemed as likely to be harmful...”. By “artefacts” Carnegie means “types
of content, aspects of the system (e.g. the way the recommender algorithm
works) and any other factors”. (emphases added)
ES: What is the harm reduction cycle?
SL: This is the core of the Carnegie model. It envisages a
repeated process of adjusting the design of intermediaries’ processes to
squeeze harm out of the system.
Carnegie has said: “Everything that happens on a social
media or messaging service is a result of corporate decisions: …” From this
perspective user behaviour is a function of the intermediaries’ processes -
processes which can be harnessed by means of the duty of care to influence, nudge
– or even fundamentally change - user behaviour.
Carnegie talk of a “Virtuous circle of harm reduction on
social media and other internet platforms. Repeat this cycle in perpetuity or
until behaviours have fundamentally changed and harm is designed out.”
There is obviously a question lurking there about the degree
to which users control their own behaviour, versus the degree to which they are
passive instruments of the platforms that they use to communicate.
ES: I sense an incipient rant.
SL: Mmm. The harm reduction cycle does prompt a vision of social
media users as a load of clothes in a washing machine: select the programme
marked harm reduction, keep cycling the process, rinse and repeat until harm is
cleaned out of the system and a sparkling new internet emerges.
But really, internet users are not a bundle of soiled
clothes to be fed into a regulator’s programmed cleaning cycle; and their words
are not dirty water to be pumped out through the waste hose.
Internet users are human beings who make decisions – good,
bad and indifferent - about what to say and do online; they are autonomous, not
automatons. It verges on an affront to human dignity to design regulation as if
people are not the cussed independent creatures that we know they are, but human
clay to be remoulded by the regulator’s iterative harm reduction algorithm. We did not
free ourselves from technology Utopianism only to layer technocratic policy Utopianism on top of it.
ES. Can system be divorced from content?
SL. It is difficult to see how, say, a recommender algorithm
can be divorced from the kind of content recommended, unless we think of recommending
content as likely to be harmful per se. Should we view echo chambers as
intrinsically harmful regardless of the content involved? Some may take that
view, perhaps the future regulator among them. Whether it is appropriate for a
regulator to be empowered to make that kind of judgement is another matter.
Realistically, a duty of care can hardly avoid being about -
at least for the most part - what steps the intermediary has to take in respect
of what kinds of content. The blue pencil - even in the guise of a broad brush
wielded by intermediaries according to Codes of Practice arranged by process – would
rest in the ultimate hand of our digital Lord Chamberlain.
ES. Surely there will be some limits on what the
regulator can do?
SL. The White Paper excluded some harms, including those
suffered by organisations. The Carnegie draft
Bill does not. It does contain a hard brake on the regulator’s power: it must
not require service providers to carry out general monitoring. That is intended
to comply with Article 15 of the Electronic Commerce Directive.
The government has also said there is no intention to include
the press. The Carnegie draft Bill attempts a translation of that into
statutory wording.
ES. Are there any other brakes?
SL. The regulator will of course be bound by human rights
principles. Those are soft rather than hard brakes: less in the nature of red
lines, more a series of hurdles to be surmounted in order to justify the
interference with the right – if the right is engaged in the first place.
The fact that a right is engaged does not mean, under
European human rights law, that the interference is always a violation. It does
mean that the interference has to be prescribed by law and justified as
necessary and proportionate.
ES. Would a duty
of care engage freedom of expression?
SL. The speech of individual end users is liable to be
suppressed, inhibited or interfered with as a consequence of the duty of care
placed on intermediaries. Since the intermediaries would be acting under legal compulsion
(even if they have some choice about how to comply), such interference is the result
of state action. That should engage the freedom of speech rights of end users. As
individual end users’ speech is affected it doesn’t matter whether corporate
intermediaries themselves have freedom of expression rights.
The right is engaged whether the interference is at the
stage of content creation, dissemination, user engagement or moderation/deletion.
These are the four points that Carnegie identifies in a paper that discusses
the compatibility of its draft Bill with fundamental freedoms.
Carnegie contemplates that tools made available to users,
such as image manipulation and filters, could be in scope of the duty of care.
It suggests that it is at least arguable that state control over access to
creation controls (paint, video cameras and such like) could be seen as an
interference. Similarly for a prohibition on deepfake tools, if they were held
to be problematic.
Carnegie suggests that it is a difficult question whether,
if we have been nudged in one direction, it is an interference to
counter-nudge. However, the question is surely not whether counter-nudging
required under a duty of care is an interference that engages the right of
freedom of expression (it is hard to see how it could not be), but whether the
interference can be justified.
ES. What if intermediaries had only to reduce the speed
at which users’ speech circulates, or limit amplification by tools such as
likes and retweets?
SL. Still an interference. People like to say that freedom
of speech is not freedom of reach, but that is just a slogan. If the state
interferes with the means by which speech is disseminated or amplified, it
engages the right of freedom of expression. Confiscating a speaker’s megaphone
at a political rally is an obvious example. Cutting off someone’s internet
connection is another. The Supreme Court of India said earlier this month, in a
judgment about the Kashmir internet shutdown:
“There is no dispute that freedom
of speech and expression includes the right to disseminate information to as
wide a section of the population as is possible. The wider range of circulation
of information or its greater impact cannot restrict the content of the right
nor can it justify its denial."
Seizing a printing press is not exempted from interference
because the publisher has the alternative of handwriting. Freedom of speech is
not just freedom to whisper.
ES: So you can’t just redefine a right so as to avoid
justifying the interference?
SL. Exactly. It is dangerous to elide the scope of the right
and justified interference; and it is easy to slip into that kind of shorthand:
“Freedom X does not amount to the right to do Y.” The right to do Y is almost
always a question of whether it is justifiable to prevent Y in particular
factual circumstances, not whether Y is generally excluded from protection.
The Carnegie fundamental freedoms paper says: “Requiring
platforms to think about the factors they take into account when prioritising
content (and the side effects of that) is unlikely to engage a user’s rights;
it is trite to say that freedom of speech does not equate to the right to
maximum reach”. However it is one thing to justify intervening
so as to reduce reach. It is quite another to argue that the user’s freedom of speech rights are
not engaged at all.
ES. So the regulator could justify restricting the provision
or use of amplification tools?
SL. If the interference is necessary in pursuance of a
legitimate public policy aim, and is proportionate to that aim, then it can be
justified. Proportionality involves a balancing exercise with other rights, and
consideration of safeguards against abuse. The more serious the interference,
the greater the justification required.
ES. That sounds a bit vague.
SL. Yes. It gets more so when we add the notion that the
state may have a positive obligation to take action to secure a right. That commonly
occurs with rights that may conflict with freedom of expression, such as
privacy or freedom of religion. We can even
end up arguing that the state is obliged to inhibit some people’s speech in
order to secure the genuine, effective right of others to enjoy their own right
of expression; to ensure that the right is not illusory; or to secure the availability
of a broad range of information.
ES. The state is required to destroy the village in order
to save it?
SL. Not quite, but increasingly
the coercive powers of the state are regarded as the means of securing freedom of
expression rather than as a threat to it.
So Carnegie questions whether removing a retweet facility is really a violation of users' rights to formulate their own opinion and express their views, or rather - to the contrary - a mechanism to support those rights by slowing them down so that they can better appreciate content, especially as regards onward sharing.
The danger with conceptualising fundamental rights as a collection of virtuous swords
jostling for position in the state’s armoury is that we lose focus on their
core role as a set of shields creating a defensive line against the excesses and
abuse of state power.
ES. So can we rely on fundamental rights to preserve the
village?
SL. The problem is that there are so few red lines. The village
can start to sink into a quagmire of competing rights that must be balanced
with each other.
The Carnegie fundamental freedoms paper well illustrates the
issue. It is a veritable morass of juxtaposed rights, culminating with the inevitable,
but not greatly illuminating, conclusion that interferences may be justified if a
fair balance is maintained between conflicting rights, and suggesting that the
state may have positive obligations to intervene. That is not a criticism of
the paper. It is how European human rights law has developed.
ES. You mentioned the ‘prescribed by law’ requirement.
How does that apply?
SL. ‘Prescribed by law’ is the first step in assessing Convention
compatibility of an interference. If you don’t pass that step you don’t move on
to legitimate objective, necessity and proportionality. Prescribed by law means
not just issued by the legislature and publicly accessible, but having the
quality of law.
Quality of law is an articulation of the rule of law.
Restrictions must be framed sufficiently clearly and precisely that someone
can, in advance, know with reasonable certainty whether their conduct is liable
to have legal consequences. In the context of the duty of care, the legal
consequence is that a user’s speech is liable to be the subject of preventive,
inhibiting or other action by a platform operator carrying out its duty.
Quality of law is a particular concern for discretionary
state powers, which by their nature are liable to be exercised ad hoc and
arbitrarily in the absence of clear rules.
ES. Discretionary power would be relevant to the
regulator, then.
SL. Yes. As we have seen, the proposed regulator would have very
wide discretion to decide what constitutes harm, and then what steps
intermediaries should take to reduce or prevent risk of harm. You can build in consultation and 'have regard' obligations, but that doesn't change the nature of the task.
ES. Couldn’t the quality of law gap be filled by the
regulator’s Codes of Practice?
SL. It is certainly possible for non-statutory material such
as Codes of Practice to be taken into account in assessing certainty and
precision of the law. Some would say that the ability of the regulator to fill
the gap simply highlights the extent to which the draft Bill would delegate
power over individual speech to the regulator.
ES. Isn’t that a doctrinaire rather than a legal
objection?
SL. Partly so. However, in the US ACLU v Reno case in
the early 1990s the US courts took a medium-specific approach, holding that the
kinds of restriction that might be justified for broadcast were not so for
individual speech on the internet. So the appropriateness of the very model of
discretionary broadcast regulation could come into question when applied to
individual speech. It bears repeating that the duty of care affects the speech
of individuals.
ES. Carnegie says that its emerging model would ‘champion
the rule of law online’.
SL. The essence of the rule of law is “the restriction of
the arbitrary exercise of power by subordinating it to well-defined and
established laws” (OED). That is characteristic of the general law, but not of
discretionary regulation.
If anything, regulation by regulator presents a challenge to
the rule of law. The raison d’etre is to maximise the power, flexibility and
discretion of the regulator. Some may
think that is a good approach. But good approach or bad, champion the rule of
law it does not.
ES. Thank you.
No comments:
Post a Comment
Note: only a member of this blog may post a comment.