...

The Myth of the Superuser: Fear, Risk, and Harm Online

by user

on
Category: Documents
41

views

Report

Comments

Transcript

The Myth of the Superuser: Fear, Risk, and Harm Online
The Myth of the Superuser:
Fear, Risk, and Harm Online
Paul Ohm∗
Fear of the powerful computer user, the “Superuser,” dominates debates
about online conflict. He is a mythic figure: difficult to find, immune to
technological constraints, and aware of legal loopholes. Policymakers,
fearful of his power, too often overreact by passing overbroad, ambiguous
laws intended to ensnare the Superuser but which are instead used against
inculpable, ordinary users. This response is unwarranted because the
Superuser is often a marginal figure whose power has been greatly
exaggerated.
The exaggerated focus on the Superuser reveals a pathological
characteristic of the study of power, crime, and security online, which
springs from a widely held fear of the Internet. Building on the social
science fear literature, this Article challenges the conventional wisdom and
∗
Associate Professor of Law and Telecommunications, University of Colorado
Law School. Thanks to Tim Wu, Orin Kerr, Phil Weiser, Julie Cohen, Pierre Schlag,
Brett Frischmann, Victor Fleischer, Miranda Fleischer, Viva Moffat, and Jean Camp
for their helpful comments. Thanks also for suggestions and support from
participants in the University of Colorado Law School Faculty Workshop, including
Clare Huntington, Nestor Davidson, Scott Peppett, Mimi Wesson, Amy Schmitz, Sarah
Krakoff, and Brad Bernthal; the Intellectual Property Scholars Conference; and the
Telecommunications Policy Research Conference. Thanks also to Todd Blair and
Michael Beylkin for outstanding research assistance.
1327
1328
University of California, Davis
[Vol. 41:1327
standard assumptions about the role of experts. Unlike dispassionate
experts in other fields, computer experts are as susceptible as laypeople to
exaggerate the power of the Superuser.
The experts in computer security and Internet law have failed to deliver
us from fear, resulting in overbroad prohibitions, harms to civil liberties,
wasted law enforcement resources, and misallocated economic investment.
This Article urges policymakers and partisans to stop using tropes of fear,
calls for better empirical work on the probability of online harm, and
proposes an Anti-Precautionary Principle — a presumption against new
laws designed to stop the Superuser.
TABLE OF CONTENTS
INTRODUCTION ................................................................................. 1329
I. THE SUPERUSER AND THE MYTH ............................................. 1332
A. The Superuser and the Myth Defined ............................... 1333
1. The Superuser........................................................... 1333
2. The Myth of the Superuser....................................... 1335
B. The Superuser and Online Conflict .................................. 1337
1. Music and Movie Piracy and DRM........................... 1338
2. Computer Security and Unauthorized Access ......... 1339
3. Surveillance .............................................................. 1340
C. The Myth of the Superuser in the Noncomputer World..... 1340
D. Dispelling the Myth of the Superuser ............................... 1342
1. Myth-Dispelling Studies and Statistics..................... 1343
2. Myth-Dispelling Anecdotes ...................................... 1345
3. Obvious Overstatements of the Risk........................ 1347
4. The Scope of the Claim ............................................ 1348
II. HARMS OF THE MYTH ............................................................. 1348
A. Overbroad Laws .............................................................. 1348
B. Unduly Invasive Search and Seizure ................................ 1353
C. Guilt by Association ........................................................ 1357
D. Wasted Investigatory Resources ....................................... 1358
E. The Myth and Legal Scholarship...................................... 1359
III. MYTH ORIGINS AND PERPETUATION ....................................... 1362
A. Myth Origins: Fear ......................................................... 1363
1. Fear ........................................................................... 1363
a. The Effects of Fear ............................................... 1363
b. Fear of the Internet............................................... 1364
2. Superusers in the Media ........................................... 1367
B. Myth Origins: Technology .............................................. 1370
1. Malleability, or the Care and Feeding
of Superusers ............................................................ 1370
2008]
The Myth of the Superuser
1329
2. Misunderstanding “Code Is Law” ............................ 1372
3. The Truth About Code: Constraints, Not Endless
Possibility ................................................................. 1375
C. Myth Perpetuation: The Failure of Expertise................... 1377
1. Pervasive Secrecy...................................................... 1378
2. Everybody Is an Expert ............................................ 1379
3. Self-Interest............................................................... 1381
4. The Need for Interdisciplinary Work ...................... 1383
IV. PRESCRIPTIONS FOR MYTH AVOIDANCE .................................. 1384
A. Superuser Harm Is Not Global Warming ......................... 1384
B. Toward a New Empirical Foundation for Online
Debates............................................................................ 1385
1. A Moratorium on Suspect Rhetoric.......................... 1385
a. Mere Possibilities ................................................. 1385
b. Secondhand Information....................................... 1386
c. Secrecy................................................................. 1386
d. Undetectable Power.............................................. 1388
2. Improving the Empirical Foundation: Toward
Better Fact-Finding................................................... 1389
a. Better Using Facts ................................................ 1390
b. Using Better Facts ................................................ 1391
c. Discovering New Facts ......................................... 1393
C. The Anti-Precautionary Principle .................................... 1393
D. Two Additional Difficulties .............................................. 1396
1. Regulating Actual Superusers................................... 1396
2. Regulating Prometheus: The Problem of Script
Kiddies...................................................................... 1399
CONCLUSION..................................................................................... 1401
INTRODUCTION
In 1992, a hacker accessed a computerized telephone switch in
Worcester, Massachusetts.1 He shut off phone service to the local
airport and disabled the runway landing lights.2 Over the past decade,
a young Norwegian man named Jon Johansen has rankled the movie
and music industries by repeatedly picking their “unbreakable” copyprotection software locks.3 In 2001, a Russian hacker named Alexy
1
2
3
See infra text accompanying notes 40-43.
See infra text accompanying notes 40-43.
See infra text accompanying note 36.
1330
University of California, Davis
[Vol. 41:1327
Ivanov broke into computer systems across America, stealing
thousands of credit card numbers.4
These three people have done things with computers that most
people could not even begin imagining how to do. They are people
locked in struggles against opposing factions who regard them and the
power they wield as grave threats. The Department of Homeland
Security worries about the airport hacker; content owners fear the lock
picker; and privacy advocates and business executives fret over the
credit card number thief. A mythology has arisen around these people
and countless others like them, whose stories have been repeatedly
retold in the news, the halls of Congress, and the pages of law reviews.
These stories could contribute usefully to debates about important
conflicts, such as computer crime, digital rights management
(“DRM”),5 and identity theft, if they were cited for what they were:
interesting anecdotes that provide a window into the empirical realities
of online conflict. Instead, these stories subsume the debates and
substitute for a more meaningful empirical inquiry. The storytellers’
pervasive attitude is, “We don’t need to probe too deeply into the
empirical nature of power in these conflicts because these stories tell us
all we need to know.” Hackers can kill airline passengers, DRM is
inherently flawed, and computer criminals steal identities by the
thousands. Through the spread of myth in a cluttered rhetorical
landscape, these fears become not merely possible, but inevitable.
Storytelling is epidemic in debates about online disputes. The
dominant rhetorical trope is the myth of power grounded in fears of
the Internet. The myth infects these debates, leading policymakers to
harmful, inefficient, and unwarranted responses because the myth is
usually exaggerated and often untrue.
Most Internet users are unsophisticated, exercising limited power
and finding themselves restricted by technological constraints, while a
minority have great power and can bypass such constraints. This
Article focuses on the powerful user — the Superuser. He (always he)
is a mythical figure: difficult to find, expensive to catch, able to
circumvent any technological constraint, and aware of every legal
loophole.
4
See infra text accompanying notes 75-76.
Digital rights management refers to technologies used to control access to
content, often copyrighted content. See infra Part I.B.1.
5
2008]
The Myth of the Superuser
1331
In this Article, I argue that there is too much focus on the Superuser.6
For most online conflicts, the Superuser likely plays a very small role.
To prove this general point, I focus on three of the most pressing and
contested ongoing online controversies: DRM, unauthorized access to
computers, and government search and surveillance of computers and
networks. I revisit these important battlegrounds throughout the
Article to demonstrate how Superuser rhetoric has distorted debate and
driven policy, despite the absence of empirical evidence that Superusers
play an important role on these battlegrounds and despite the presence
of some evidence to the contrary.
Further, I argue Superuser stories spring from a widely held fear of
the Internet. The link between fear and the Superuser exemplifies
some general conclusions from behavioral economics and cognitive
psychology about fear and how it causes people to exaggerate risk by
triggering biases and heuristic shortcuts.
This Article’s examination of the Myth of the Superuser extends the
pre-existing literature by questioning assumptions about the role of
experts. Unlike their counterparts in other fields, experts in network
security and computer crime usually make the same errors in
6
To date, no one has taken a comprehensive, theoretical, and systematic look at
the harm that results from over-attention to and exaggeration of the powerful computer
user. A few scholars have discussed this idea in passing. For example, Professor
Lawrence Lessig has distinguished between “hackers” and “the rest of us” and has
argued that the existence of the former should not stop us from trying to solve
problems that primarily affect the latter. Lawrence Lessig, Constitution and Code, 27
CUMB. L. REV. 1, 3 (1997) (“I don’t choose whether to obey the structures that [code]
establishes — hackers might, but hackers are special. . . . For the rest of us, life in
cyberspace is subject to the code of cyberspace, just as life in real space is subject to the
code of real space”); Lawrence Lessig, The Constitution of Code: Limitations on ChoiceBased Critiques of Cyberspace Regulation, 5 COMMLAW CONSPECTUS 181, 184 (1997);
Lawrence Lessig, Reading the Constitution in Cyberspace, 45 EMORY L.J. 869, 896 n.80
(1996) [hereinafter Lessig, Reading]; Lawrence Lessig, The Zones of Cyberspace, 48
STAN. L. REV. 1403, 1408 n.17 (1996) (explaining that “what hackers do doesn’t define
what the effect of law as code is on the balance of the non-hacker public”).
Professor Timothy Wu has advanced a similar theme. See Timothy Wu, ApplicationCentered Internet Analysis, 85 VA. L. REV. 1163, 1203-04 (1999) (“[E]xpert users suffer
least and benefit most from an unregulated Internet. . . . [T]o stick everyone with the
constitution of the expert user may, in the long run, prove the inexpert move, as it
may do more to close out the Internet than flexibility ever would.”); see also JACK
GOLDSMITH & TIM WU, WHO CONTROLS THE INTERNET? ILLUSIONS OF A BORDERLESS
WORLD 123 (2006) (predicting world in which minority users “with all the time and
expertise” continue to download free music while rest use legitimate pay sites).
None of these scholars has explored the exaggerated Myth of the Superuser more
deeply, situated these observations within the sociological literature of fear, examined
the negative effects that result from relying on the Myth, or provided detailed
prescriptions for dealing with these effects.
1332
University of California, Davis
[Vol. 41:1327
judgment as laypeople. These experts spout blusterous exaggeration
about online threats rather than acting as calm, measured assessors of
truth. In part, these experts have been led astray by incorrect
assumptions about the malleability of code and misinterpretations of
Lessig’s observations about code and law.7
I define the Superuser and the Myth of the Superuser (“Myth”) in
Part I and establish that the Myth is often an exaggeration. In Part II, I
show how undue attention to the Myth of the Superuser has been
harmful to civil liberties, efficient and effective law enforcement, and
sensible Internet regulation. In Part III, I tie the persistence of the
Myth to the widely held fear of the Internet. Faced with the general
public’s fear, experts have abandoned their responsibility to be
dispassionate truth seekers and instead have engaged in the rhetoric of
myth and exaggeration.
Finally, in Part IV, I offer prescriptions for lawmakers, judges, and
scholars to blunt the use and effect of the Myth. First, I challenge a set
of commonly used rhetorical tools that are the hallmark of Superuser
myth-telling. Second, I call for a new approach for counting
Superusers. Finally, I urge regulators to adopt an Anti-Precautionary
Principle — in the absence of any empirical proof about an online
threat or harm, legislators should refrain from regulating anew.
In the ongoing dialogue about how best to regulate virtual,
constructed spaces, we find ourselves awash in metaphor, analogy,
and myth. These myths carry great weight, repeated by those with the
trappings of authority and never challenged for accuracy or even
plausibility. By calling into question the dominant myth, the Myth of
the Superuser, I try to restore some of what we have lost: constructive
debate and carefully reasoned regulation.
I.
THE SUPERUSER AND THE MYTH
The rhetorical devices I define in this Part are not uncommon; they
are pervasive. Every important debate about online conflict leads
eventually to one or both sides making claims about the Superuser.8
7
See generally LAWRENCE LESSIG, CODE: VERSION 2.0 (2d ed. 2006) (describing
how code regulates online behavior).
8
I restate it as a less-charged version of Godwin’s Law: As a debate about online
conflict grows longer, the probability of an argument involving powerful computer
users approaches one. MIKE GODWIN, CYBER RIGHTS 48 (1998) (“As an online
discussion grows longer, the probability of a comparison involving Nazis or Hitler
approaches one.”). I would call it “Ohm’s Law,” but that is already taken.
2008]
The Myth of the Superuser
1333
A. The Superuser and the Myth Defined
1.
The Superuser
A few terms must be defined. As used in this Article, “power” is the
ability to control or change computers or networks. If computer users
are rank-ordered by the amount of power they possess, the “ordinary
user” is, roughly speaking, the user in the middle.
A “Superuser” is a computer user who possesses power that the
ordinary user does not.9 He can control or change computers and
networks in extraordinary ways. Superusers tend to have more time,
practice, knowledge, or access to tools than ordinary users.10 Tools
play a particularly important and complex role. Superusers often use
sophisticated computer programs (sometimes created by them, often
created by others) to gain power.11
9
A note on terminology: the word “Superuser” has not been used before in legal
scholarship with precisely this meaning. In various versions of UNIX and UNIX-like
Operating Systems, “superuser” (lowercase “s”) is the name given to the user account
that a system administrator can use to make almost any change to the system. See EVI
NEMETH ET AL., UNIX SYSTEM ADMINISTRATION HANDBOOK 39 (3d ed. 2001). This is also
known as the “root” account. Id. at 37. Some commentators have used this meaning
of superuser. See, e.g., Richard W. Downing, Shoring Up the Weakest Link: What
Lawmakers Around the World Need to Consider in Developing Comprehensive Laws to
Combat Cybercrime, 43 COLUM. J. TRANSNAT’L L. 705, 721 (2005) (discussing hacker
who “gains ‘root level’ access, also known as ‘superuser’ status”).
My meaning is different, and in this Article I capitalize “Superuser” to distinguish it
from the prior, UNIX definition. Several technically minded readers have criticized my
decision to give a new meaning to this word, accusing me of muddying an otherwise
clear term or, less charitably, suggesting that it betrays a lack of technical knowledge or
acumen. See the various comments to Slashdot, http://it.slashdot.org/article.pl?sid=
07/04/11/1952247 (last visited Apr. 3, 2008) and The Volokh Conspiracy,
http://www.volokh.com/posts/1176127892.shtml (last visited Apr. 3, 2008).
I decided to use the term Superuser despite these criticisms because it concisely
captures the idea of “power without necessary malice” that alternatives like “hacker”
or the neologism, “superhacker” do not convey. Also, although many computer
experts — particularly UNIX specialists — know the term’s other meaning, most
noncomputer specialists, including policymakers, do not. Finally, even for those who
know the earlier, UNIX meaning of the term, it should be easy to understand which of
the two meanings is meant from context.
10
Cf. Paul Graham, Great Hackers, July 2004, http://www.paulgraham.com/
gh.html (describing skills required to be hacker); Eric Steven Raymond, How to
Become a Hacker, http://catb.org/~esr/faqs/hacker-howto.html (last updated Jan. 8,
2008) (same). Often, they have more than one of these, but they need not have all of
them.
11
CHRIS PROSISE & KEVIN MANDIA, INCIDENT RESPONSE & COMPUTER FORENSICS 385414 (2d ed. 2003) (entitling chapter “Investigating Hacker Tools”); Graham, supra
1334
University of California, Davis
[Vol. 41:1327
But once a tool becomes widely used, its users are considered
ordinary users and not Superusers. For example, in the late 1990s,
Napster users could browse and copy the music collections of millions
of others, yet they were not Superusers.12 Because many people used
their software, these people were merely empowered ordinary users.
In other words, the term Superuser is a relative one. Having power —
the ability to effect change — is necessary but not sufficient to be a
Superuser.
The term can come and go with time. A person with power X is a
Superuser only as long as the percentage of people with X is small.
Usually, that percentage increases over time, and once X becomes
accessible to many people, the Superuser designation disappears.
Consider audio compact disc ripping. Not long ago, when CDs and
the computer drives that could read them were both new and scarce,
few people (all Superusers) could copy the music from a CD onto a
computer or “rip” the CD.13 It was not long, however, before
Superusers packaged this power into functional but unpolished
computer programs used by people not considered Superusers.14
Today, the ability to rip a CD is no longer a Superuser power. This is
the result of programmers making ripping programs both easier to
obtain15 and use.16
note 10 (“Like all craftsmen, hackers like good tools.”).
12
See A&M Records v. Napster, Inc., 239 F.3d 1004, 1010-13 (9th Cir. 2001)
(describing Napster).
13
See Mike Snider, Microsoft, Macrovision Join to Halt CD ‘Ripping,’ USA TODAY,
Apr. 24, 2003, at 5D.
14
When Superusers package their power into tools, the empowered but ignorant
users are often called script kiddies. See SIMSON GARFINKEL, WEB SECURITY, PRIVACY AND
COMMERCE 401 (2d ed. 2002). We will consider the script kiddie again in Part IV.D.2.
15
CD ripping software now comes bundled, for example, with Windows
XP/Windows Media Player and Mac OS X/iTunes. See Apple Inc., iTunes 7: How to
Improve Performance While Burning or Ripping CDs, http://docs.info.apple.com/
article.html?artnum=304410 (last visited Apr. 3, 2008); Windows Media, Ripping CDs
in Windows Media Player,
http://www.microsoft.com/windows/windowsmedia/
knowledgecenter/mediaadvice/0080.mspx (last visited Apr. 3, 2008).
16
There are countless other examples of the passage of time redefining Superuser
power. Consider photo sharing. A decade ago, to share photos on the web, you had
to scan physical prints into digital files, upload the files using the file transfer
protocol, write (without the assistance of any specialized tools) a web page containing
those photos, and then email the URL to your friends and relatives. A little less than a
decade ago, you could use an early-model digital camera and a web-hosting service
like Geocities to develop a photo gallery using better but still clunky tools. Today, an
account with Flickr or Kodak Gallery accomplishes the same goal in much less time
with significantly better results.
Further, to send an anonymous email in the early 1990s, you had to issue a series of
2008]
The Myth of the Superuser
1335
Not all Superusers act with malice or try to cause harm. There are
good Superusers, bad Superusers, and morally neutral Superusers. As
the CD ripping example demonstrates, sometimes the morality of a
Superuser’s power is in the eye of the beholder. Most people consider
CD ripping a morally neutral act, but some copyright owners may see
it as a harmful, negative act.17
2.
The Myth of the Superuser
The “Myth of the Superuser” is the belief that an online conflict
cannot be resolved without finding a way to neutralize Superusers. As
I explore more fully in Part I.D, the Myth is flawed. Superusers are
often inconsequential because they are uncommon or unable to cause
great change. Thus, I argue solutions targeted at ordinary users are
good enough.
The Myth also refers to any argument that invokes the Superuser to
support or oppose a proposed solution to an online conflict. Consider
Congress’s rhetoric supporting amendments to the Computer Fraud
and Abuse Act (“CFAA”), the federal law criminalizing computer
hacking and trespass, codified in § 1030 of title 18 of the U.S. Code.18
Specifically, consider the 1996 Senate Committee Report19 (“1996
precise commands (which complied with the SMTP email protocol) to an email
server. In the late 1990s, anonymous remailers in foreign countries would strip
identifying information from incoming messages and forward them onto their
destination. Today, setting up an account at Yahoo! Mail or Gmail quickly enables
one to send pseudonymous email messages.
17
See Posting of Fred von Lohmann to Deeplinks Blog, http://www.eff.org/
deeplinks/archives/004409.php (Feb. 15, 2006) (quoting RIAA in regulatory filing as
saying, “creating a back-up copy of a music CD is not a non-infringing use”).
18
Congress adopted § 1030 in 1984. See Counterfeit Access Device and
Computer Fraud and Abuse Act of 1984, tit. II, Pub. L. No. 98-473, § 2101, 98 Stat.
1837, 2190. It has made at least five major revisions to § 1030 since then. See, e.g.,
Homeland Security Act of 2002, tit. II, Pub. L. No. 107-296, sec. 225, § 1030, 116 Stat.
2135; Uniting and Strengthening America by Providing Appropriate Tools Required to
Intercept and Obstruct Terrorism (USA PATRIOT) Act of 2001, tit. V, Pub. L. No.
107-56, sec. 506, § 1030, 115 Stat. 272, 278; Economic Espionage Act (EEA) of 1996,
Pub. L. No. 104-294, 110 Stat. 3488, 3491-95; Violent Crime Control and Law
Enforcement Act of 1994, tit. XXIX, Pub. L. No. 103-322, sec. 290001, § 1030, 108
Stat. 1796; Computer Fraud and Abuse Act of 1986, Pub. L. No. 99-474, sec. 2, §
1030, 100 Stat. 1213. These frequent amendments are surprising, given the
infrequency with which the law is used. Cf. Orin Kerr, Lifting the “Fog” of Internet
Surveillance: How a Suppression Remedy Would Change Computer Crime Law, 54
HASTINGS L.J. 805, 823-24 (2003) (noting relative infrequency with which computer
privacy provisions of 18 U.S.C. § 2701 are invoked).
19
See generally S. REP. NO. 104-357 (1996).
1336
University of California, Davis
[Vol. 41:1327
Report”) accompanying broad amendments to the CFAA.20 This
report is a model of Superuser storytelling, full of questionable
rhetorical devices typical of the Myth.21
First, Superuser stories are simultaneously detailed and vague. They
richly describe events but omit the names, dates, and places needed to
corroborate the story.22 The 1996 Report is full of anecdotes about
nefarious Superusers. For example, the Committee asserted that
“[h]ackers . . . have broken into . . . supercomputers for the purpose
of running password cracking programs.”23 The hackers and victims
are anonymous, and the authors never say whether these mythical
criminals were caught or whether gaps in the law hindered attempts to
prosecute them.
Second, Superuser storytellers summarize trends using vague
adverbs like “often,” “usually,” and “frequently.”24
To justify
broadening the scope of a crime under the CFAA, the Committee
asserted that “intruders often alter existing log-on programs so that
20
National Information Infrastructure Protection Act (NIIPA) of 1996, tit. II, Pub.
L. 104-294, § 201, 110 Stat. 3488, 3491-94 (passing NIIPA as Title II of EEA).
21
The Economic Espionage Act (“EEA”), which criminalizes certain trade secrets
thefts (codified at 18 U.S.C. §§ 1831-1839 (2000 & Supp. V 2005)), is another
example of a computer crime law whose prosecution record does not seem to match
the hype and rhetoric used by those who urged the creation of the law. See Joseph F.
Savage, Jr., Matthew A. Martel, & Marc J. Zwillinger, Trade Secrets Conflicting Views of
the Economic Espionage Act, 15 CRIM. JUST. 10, 11-12 (2000) (contrasting foreign
government prohibitions of Act with domestic trade secret theft provision that has
been frequently used). Consider the views of two defense attorneys (writing with a
former federal prosecutor) about the EEA:
The congressional debates contained dire accounts of foreign governments
pilfering America’s trade secrets. Simply put, the EEA was couched in terms
of national security. . . . Because of this original focus, one might surmise
that the DOJ would immediately give priority to enforcement procedures
involving foreign spies. Instead, not one prosecution has occurred enforcing
the foreign espionage provisions of the EEA. Not a single one. The
conclusion seems inescapable: The foreign economic espionage law either
was not necessary or there is a real and ongoing problem that is not being
addressed.
Id. at 11.
22
Cf. Michael Levi, “Between the Risk and the Reality Falls the Shadow”: Evidence
and Urban Legends in Computer Fraud (with Apologies to T.S. Eliot), in CRIME AND THE
INTERNET 44, 46 (David S. Wall ed., 2001) (discussing computer security experts’
claims that events happened, but discussing those claims without pertinent details in
service of client confidentiality).
23
S. REP. NO. 104-357, at 9.
24
See, e.g., id. at 11.
2008]
The Myth of the Superuser
1337
user passwords are copied to a file that the hackers can retrieve
later.”25 Again, the Committee provided no other details.
Third, Superuser storytellers pass the buck, parroting back what
others have told them. To justify a new crime to address the allegedly
pressing problem of people making threats against computer
systems,26 the Committee said, “According to the Department of
Justice [(“DOJ”)], threats have been made against computer systems
in several instances.”27 The DOJ’s authorship of the vague claim is the
only proof of its relevance and veracity.
Fourth, Superuser stories are often hypothetical. To justify the same
computer threat provision, the Committee mused, “One can imagine
situations in which hackers penetrate a system, encrypt a database and
then demand money for the decoding key.”28
These are not the only strategies employed by Superuser storytellers.
They also rely on exaggeration, jargon, strained metaphors, appeals to
common sense and common knowledge, and other rhetorical sleightsof-hand.29
B. The Superuser and Online Conflict
In every online conflict, partisans wield the Myth of the Superuser
like a rhetorical bludgeon.30 The Superuser has become a form of
what sociologist Stanley Cohen termed the “folk devil”:31 members of
25
Id. (emphasis added).
See 18 U.S.C. § 1030(a)(7) (2000) (making it crime to, “with intent [] extort
from any person any money or other thing of value, transmit[] in interstate or foreign
commerce any communication containing any threat to cause damage to a protected
computer”).
27
S. REP. NO. 104-357, at 12.
28
Id.
29
See PAUL A. TAYLOR, HACKERS: CRIME IN THE DIGITAL SUBLIME, at xii (1999) (“The
rhetoric of the digital sublime describes the particularly high levels of hyperbole that
seem to surround computer-based technologies.”) (emphasis omitted); MAJID YAR,
CYBERCRIME AND SOCIETY 24-25 (2006) (describing “hyperbole” of “official
pronouncements on hacking”).
30
The phrase “online conflict” is purposefully broad and used to describe any of
the disputes, lawsuits, or debates that have arisen as a result of the emergence of the
Internet. In the pages that follow, the necessarily vague term is used somewhat
inconsistently, on occasion to describe very specific clashes between well-defined
stakeholders, while on others to describe more amorphous disagreements between less
clearly delineated sides.
31
STANLEY COHEN, FOLK DEVILS AND MORAL PANICS: THE CREATION OF THE MODS
AND ROCKERS 9-11 (1972); cf. Carol Sanger, Infant Safe Haven Laws: Legislating in the
Culture of Life, 106 COLUM. L. REV. 753, 782 (2006) (discussing Cohen’s theories of
folk devils and moral panics in context of infanticide).
26
1338
University of California, Davis
[Vol. 41:1327
society seen as dangerous or deviant, blamed for many of society’s
ills.32 The trope arises in every single branch of Internet law,
including intellectual property, computer crime, information privacy,
information security, Internet governance, telecommunications,
innovation policy, First Amendment law, and jurisdiction.
This Article focuses on three important conflicts: DRM, computer
crime, and computer search and seizure. These represent a broad
cross- section of policy concerns and styles of debate and argument.
These also raise three different sets of institutional concerns, because
these debates have been targeted, primarily and respectively, at
scholars, legislators, and judges.
1.
Music and Movie Piracy and DRM
DRM systems allow content owners to control what other people
can do with data.33 A fierce debate over DRM rages, focusing largely
on whether DRM systems should stand or fall on their own technical
merits, or if instead they should be bolstered by laws that make it
illegal to circumvent DRM (i.e., pick the locks) or to teach others to
do the same.34 The Superuser looms large in the debate because he
can pick locks that ordinary users cannot.
Consider again Jon Johansen. Apple’s iTunes Music Store sells
songs protected by a DRM technology called FairPlay, which lets
purchasers listen to the music they have bought only in authorized
ways and only on authorized computers.35 Johansen, a Superuser par
excellence, has repeatedly created programs that can strip the
32
See John Timmer, Breaches of Personal Data: Blaming the Myth and Punishing the
Victim, ARS TECHNICA, Mar. 14, 2007, http://arstechnica.com/news.ars/post/20070314breaches-of-data-blaming-the-myth.html (explaining that “hackers have become the
folk devils of computer security”).
33
See generally Dan Burk, Legal and Technical Standards in Digital Rights
Management Technology, 74 FORDHAM L. REV. 537 (2005) (discussing role DRM plays
in shaping and replacing technical and legal standards); Julie E. Cohen, DRM and
Privacy, 18 BERKELEY TECH. L.J. 575 (2003) (identifying connection between DRM and
privacy protection); Randall Picker, Mistrust-Based Digital Rights Management, 5 J. ON
TELECOMM. & HIGH TECH. L. 47 (2006) (proposing methods for disincentivizing DRM
circumvention by tying copies of works to purchaser).
34
See Julie E. Cohen, Some Reflections on Copyright Management Systems and Laws
Designed to Protect Them, 12 BERKELEY TECH. L.J. 161, 162-63 (1997); Pamela
Samuelson, Intellectual Property and the Digital Economy: Why the Anti-Circumvention
Regulations Need To Be Revised, 14 BERKELEY TECH. L.J. 519, 522-23 (1999).
35
Steve Jobs, CEO of Apple, has publicly called the use of DRM for music into
question. Steve Jobs, CEO, Apple Inc., Thoughts on Music (Feb. 6, 2007),
http://www.apple.com/hotnews/thoughtsonmusic/.
2008]
The Myth of the Superuser
1339
protection from a FairPlay-protected song, empowering users to access
iTunes-purchased music in ways that Apple has tried to forbid.36
2.
Computer Security and Unauthorized Access
Computer security systems protect Internet-connected computers
from unauthorized intruders.37 These security systems are complex.38
Because of this complexity, software tends to be riddled with
vulnerabilities, many of which can be exploited by Superusers to gain
unauthorized access.39
For example, in the early 1990s, a young man bypassed a
computer’s security system to gain access to a computer in Worcester,
Massachusetts that controlled an important telephone switch.40
Inadvertently, he reset the switch, disabling local phone service to the
area, which included a small, unmanned airport.41 By disabling the
switch, he made it impossible for incoming aircraft to turn on the
landing lights.42 No planes crashed, and nobody was injured. But
advocates to this day retell the story to policymakers, spinning
terrifying variations: replace the juvenile with international terrorists,
the small regional airport with O’Hare International Airport, and the
36
See Robert Levine, Unlocking the iPod, FORTUNE, Oct. 30, 2006, at 73.
See generally Robert W. Hahn & Anne Layne-Farrar, The Law and Economics of
Software Security, 30 HARV. J.L. & PUB. POL’Y 283, 285 (2006) (giving “a
comprehensive assessment of the software security issue using a law and economics
framework”).
38
Id. at 319 (discussing market pressures that lead security software to include
many features, increasing complexity of security systems).
39
Id. at 296 (“Given the complexity of programs such as these, most experts
believe that bugs and other vulnerabilities are inevitable.”).
40
Many public and private figures have retold this story, perhaps most
prominently Scott Charney in a report issued by the House of Representatives. See
Scott Charney, Transition Between Law Enforcement and National Defense, in SECURITY
IN THE INFORMATION AGE: NEW CHALLENGES, NEW STRATEGIES 52 (Robert F. Bennet ed.,
2002), available at http://www.house.gov/jec/security.pdf. At the time he authored
this report, Charney was a principal at PricewaterhouseCoopers. Immediately before
he took that position, however, he had been a long-standing career employee at the
DOJ, where he founded and served as Chief to the Department’s Computer Crime and
Intellectual Property Section. After leaving PricewaterhouseCoopers, he became
Microsoft’s Vice President for Trustworthy Computing. See Microsoft Corp.,
http://www.microsoft.com/presspass/exec/charney/default.mspx (last visited Apr. 3,
2008) (providing Scott Charney’s biography).
41
See Charney, supra note 40, at 54.
42
See id. at 54-55.
37
1340
University of California, Davis
[Vol. 41:1327
inadvertent effect with an intentional attack, and the lesson is
unmistakable.43
3.
Surveillance
The War on Terror has given birth to many stories, including the
tale of the cat-and-mouse game between competing groups of
Superusers: terrorist data hiders and government data finders.44
Government agents use complex tools and techniques to sift through
large volumes of data stored on computers and coursing through
networks.45 They would usually find the evidence they sought, we are
told, if not for Superuser data hiders and their creative and evolving
techniques.46
Versions of the following myth are told often in law enforcement
circles.47 FBI agents serve a warrant to search John Smith’s house and
seize any computers found. Amid the usual chaos of an ongoing
search, an agent notices that a computer in the corner is whirring
loudly. Later, a forensic analysis reveals that a large portion of the
computer’s hard drive has been wiped clean, supporting fears that
Smith had triggered a software “logic bomb” after learning of the
search. It is a compelling story, but I could find no documented
examples suggesting that the myth had any basis in reality.
C. The Myth of the Superuser in the Noncomputer World
Are Superusers limited to the virtual world? Are there not people in
the physical, nonvirtual world with the knowledge, time, training, or
resources to flout technical and legal constraints? Consider locks.
Very few people know how to pick them. For most of us, locks keep
43
See id. at 55 (comparing Worcester incident to potential attack on O’Hare
International Airport).
44
See infra Part II.B.
45
Orin Kerr, Searches and Seizures in a Digital World, 119 HARV. L. REV. 531, 538
(2005).
46
See id. at 546-47.
47
Books about computer forensics, for example, often describe the supposedly
widespread use of “logic bombs” or “kill switches.” See JOHN W. RITTINGHOUSE &
WILLIAM M. HANCOCK, CYBERSECURITY OPERATIONS HANDBOOK 390 (2003) (“Since most
good computer criminals are trying to avoid detection and prosecution, they will often
employ the use of logic bombs embedded within system administrative processes
commonly required during investigations.”); JOHN R. VACCA, COMPUTER FORENSICS:
COMPUTER CRIME SCENE INVESTIGATION 238 (2d ed. 2005) (“The computer
investigator . . . needs to be worried about destructive process and devices being
planted by the computer owner . . . .”).
2008]
The Myth of the Superuser
1341
us out of places we are not meant to go and secure things that we
might otherwise want to take.
In the physical world, lawmakers often exaggerate the criminal threat
to justify broad criminal prohibitions and new law enforcement
surveillance capabilities. Witness the rhetoric surrounding child
abductions or school shootings. These crimes are often said to be
epidemic despite numerous studies that conclude otherwise.48 If
Superusers and the Myth thrive offline as well as online, what is gained
by focusing solely online? Is there anything special about this context?
Despite some similarities between the online and offline uses and
abuses of power, there are many differences between the two. These
differences make the online threat of the Myth of the Superuser much
more prevalent and vexing. The Myth of the Superuser is different in
kind and degree from real world analogs. The most important
difference is the type of power. The online Superuser is said to possess
science fiction-like abilities unlike the much more constrained power
attributed to his physical world counterparts.49 Fear of unbounded
power leads policymakers to regulate online harms in much more
broad, vague, sweeping, and as I argue later, harmful ways.50
In the real world, the Myth of the Superuser is invoked less often
because countervailing voices oppose exaggerated threats, soothing
fears and debunking myths.51 Academic researchers and some in the
media, for example, use statistics to show that concerns about
48
See Diana Griego & Louis Kilzer, The Truth About Missing Children:
Exaggerated Statistics Stir National Paranoia, DENVER POST, May 12, 1985, at 1-A
(debunking, in Pulitzer Prize winning series, exaggerated fears about child
abductions); Lynnell Hancock, The School Shootings: Why Context Counts, COLUM.
JOURNALISM REV., May-June 2001, at 76 (concluding statistics support view that
public’s increased fear of school shootings was “exaggerated, fed by saturation media
coverage that is painting a distorted picture”).
49
See infra Part III.B.2 (describing nature of Superuser’s power).
50
See infra Part II. The closest real world parallel is the “Superterrorist,” an
increasingly mythologized figure whose command over the physical world approaches
the Superuser’s control of the online world. The Superterrorist is a master at evasion,
able to plan and fund complex crimes without leaving behind any tracks. Because too
much can be made of this comparison, I spend very little additional space developing
it. I believe, however, that some of the observations and prescriptions that follow
apply to the terrorism context, as well. There is one especially salient connection
between terrorism and computer crime. Many have taken to talking about the
“cyberterrorist,” who allegedly advances terror goals by attacking computer systems.
See Joshua Green, The Myth of Cyberterrorism, WASH. MONTHLY, Nov. 2002,
http://www.washingtonmonthly.com/features/2001/0211.green.html
(debunking
claims of prevalence of cyberterrorism).
51
See, e.g., sources cited supra note 48.
1342
University of California, Davis
[Vol. 41:1327
particular types of harmful behavior are overblown.52 For reasons that
I will discuss in depth, this rarely happens with online risks.53
Finally, although lawmakers sometimes fall prey to the Myth in
regulating physical space, they often do not. Consider locks again.
Criminal laws prohibiting theft and breaking and entering are still
considered effective despite the fact that some people can pick locks.54
D. Dispelling the Myth of the Superuser
British criminologist David S. Wall has remarked:
Fears, which, in the absence of reliable information to the
contrary, have been nurtured and sustained by media
sensationalism. Yet, our practical experience of the Internet is
that few of these fears have actually been reali[z]ed.
Furthermore, there is clearly emerging a body of evidence to
show that the criminal reality of the Internet is not the all
engulfing ‘cyber-tsunami’, but, like the terrestrial world, a
large range of frequently occurring small-impact crimes.55
Superusers may walk among us, but they usually do so in small
enough numbers to be safely ignored. Even though a few Superusers
can cause harm, they are so difficult to find and apprehend, so
resistant to ordinary disincentives, or constitute so small a part of the
problem that they are not worth the hunt. Of course, even a few
Superusers demand attention if they are powerful enough to account
for a significant portion of the harm. Measuring the impact of the
Superuser requires more than a head count; it must also account for
the amount of harm caused by any one Superuser. Three reasons
illustrate why the Myth is an exaggeration.
52
See infra Part I.D.1.
See infra Part III.
54
See Lessig, Reading, supra note 6, at 896 n.80 (explaining that “from the fact
that ‘hackers could break any security system,’ it no more follows that security systems
are irrelevant than it follows from the fact that ‘a locksmith can pick any lock’ that
locks are irrelevant”); Wu, supra note 6, at 1195 (asking people to “[c]onsider for a
moment the observation that a lock may be picked; interesting, no doubt, but not a
convincing demonstration that a lock cannot serve any regulating function”).
55
See CRIME AND THE INTERNET, supra note 22, at xi.
53
2008]
1.
The Myth of the Superuser
1343
Myth-Dispelling Studies and Statistics
Statistics suggest the Superuser is a myth.56 In 2006, two
researchers at the University of Washington surveyed twenty-six years
of national print and broadcast news for stories about electronic data
loss.57 In all, 550 separate incidents were studied, amounting to the
reported loss of 1.9 billion records, or nine private records for every
adult living in the United States.58 The researchers tested the
conventional wisdom that hackers were mostly to blame for the loss of
personal data.59 To the contrary, “for the period between 2000 and
2006, 31% of the incidents were about a breach caused by a hacker,
8% of the incidents involved an unspecified breach, and 61% of the
incidents involved different kinds of organizational culpability.”60
Organizational culpability included cases involving accidental records
release, employee misconduct, misplaced backup tapes, and stolen
laptops.61
As another example, British sociologist Michael Levi focused on
studies of victimization rates for computer crime. He concluded that
they showed much less victimization than contemporary media
accounts had suggested.62 Granted, many of these studies are by now
a decade old or older. But at least when Levi drew his conclusions, the
numbers did not seem to square with the public rhetoric.
Levi noted “international surveys in 1996 and 1998 [by Ernst &
Young] . . . turned up very few cases of reported or unreported
computer frauds.”63 He also cited these older studies to show that
although more than two-thirds of executives in large private sector
56
Part IV discusses better available and more reliable sources for statistics.
Kris Erickson & Philip N. Howard, A Case of Mistaken Identity? News Accounts
of Hacker and Organizational Responsibility for Compromised Digital Records, 19802006, 12 J. COMPUTER MEDIATED COMM. 1, 3 (2007). The studies included all stories
from 1980 to 2006, using both the LexisNexis and Proquest databases.
58
Id. at 12-13.
59
Id. at 3 (stating that “the campaign against hackers has successfully cast them as
the primary culprits to blame for insecurity in cyberspace”).
60
Id. at 17. Even if the entire 8% of the unattributed breaches were the result of
hackers, hackers would still account for only 39% of the reported incidents.
61
Id. at 17-18. Erickson and Howard note, however, that hackers accounted for a
vast majority of the number of stolen records, but only because one incident — data
theft from the Acxiom Corporation that led to a criminal conviction — involved the
loss of 1.6 billion records. Id. at 18. If that single incident is removed from the data
set, hackers caused 32% of lost data, organizational behavior caused 48%, and 20%
remains unattributed. Id.
62
Levi, supra note 22, at 51-55.
63
Id. at 53 (citing 1996 and 1998 studies by Ernst & Young).
57
1344
University of California, Davis
[Vol. 41:1327
companies felt computer viruses and hacking were serious concerns,
only 11% had experienced either and only 5% had experienced both.64
A U.K. government-run audit similarly found that the number of
entities reporting information technology fraud fell from 10% in 1994
to 8% in 1997.65
Further, statistics suggesting that the Superuser is a potent force are
often rebuttable. The statistics most cited about computer security
incidents are those reported in the annual Computer Security Institute
(“CSI”) / FBI Computer Crime and Security Survey (“Survey”), a new
version of which has been released every year for over a decade.66 The
Survey may be the media’s pre-eminent source for statistics about
online harm.67 When they cite the Survey, the media often report the
results in breathless tones.68 Dozens of law review articles and student
notes have also cited the Survey.69
64
Id.
Id. at 51-52 (citing results from three reports by Audit Commission). Levi
noted that despite this drop, the average cost to each victim rose and the percentage of
respondents reporting other “IT abuse” rose during the same period. Id.
66
LAWRENCE A. GORDON ET AL., 2006 COMPUTER SECURITY INSTITUTE / FBI
COMPUTER CRIME AND SECURITY SURVEY, available at http://i.cmpnet.com/gocsi/db_area/
pdfs/fbi/FBI2006.pdf.
67
Ira Winkler, Opinion, Investigating the FBI’s ‘Invalid’ Security Survey,
SEARCHSECURITY.COM, Jan. 19, 2006, http://searchsecurity.techtarget.com/columnItem/
0,294698,sid14_gci1159992,00.html (stating that Survey is “often referred to as ‘the
most quoted stud[y] in the field’”).
68
The release of the 2006 Survey prompted this reporting: “Virus attacks,
unauthorized access to computer systems and other forms of cybercrime account for up
to 75% of the financial losses at U.S. companies.” Tim Scannell, Computer Crime and the
Bottom Line, INTERNETNEWS.COM, July 20, 2006, http://www.internetnews.com/
stats/article.php/3621236.
69
A search of the Westlaw JLR database on January 24, 2008, for the string “CSI”
/S “COMPUTER CRIME” /S “SURVEY” returned 57 hits. See, e.g., Lilian Edwards,
Dawn of the Death of Distributed Denial of Service: How to Kill Zombies, 24 CARDOZO
ARTS & ENT. L.J. 23, 31-32 (2006) (citing Survey’s reported incidence of Denial of
Service attacks); Michael L. Rustad, The Negligent Enablement of Trade Secret
Misappropriation, 22 SANTA CLARA COMPUTER & HIGH TECH. L.J. 455, 521 (2006)
(citing Survey’s finding that 70% of computer intrusions are traceable to Internet
connection); Bruce P. Smith, Hacking, Poaching, and Counterattacking: Digital
Counterstrikes and the Contours of Self-Help, 1 J.L. ECON. & POL’Y 171, 171 n.1 (2005)
(citing Survey’s statistics about spread of antivirus and firewall software); Debra Wong
Yang & Brian M. Hoffstadt, Countering the Cyber-Crime Threat, 43 AM. CRIM. L. REV.
201, 201 n.5 (2006) (citing Survey’s report of $130 million damages from
unauthorized use of computers).
65
2008]
The Myth of the Superuser
1345
The Survey’s methodology, however, is entirely suspect.70 First, the
Survey is sent only to members of the CSI.71 Second, in 2006, the
Survey was returned by a mere 12% of those who had received it.72
Although past editions of the Survey bore a disclaimer that the results
were not statistically significant, recent versions do not contain the
same disclaimer.73 Although these methodological shortcomings do
not point toward a systematic inflation or deflation of results, they do
call into question the authority with which the Survey is usually
regarded and cited. Further, even if the Survey results are given any
credence, they show that both the level of computer security incidents
and money spent on response have been decreasing steadily for the
past four years.74
2.
Myth-Dispelling Anecdotes
The second reason to doubt claims about Superuser power is that
anecdotally, some online crimes seem to be committed by ordinary
users much more often than by Superusers. Consider the growing
problems of data breach and identity theft. Data breachers are often
portrayed as genius hackers who break into computers to steal
thousands of credit cards.75 Although there are criminals who fit this
profile, the police increasingly focus on people who obtain personal
data in much more mundane, non-Superuser ways.76 For example,
70
See Winkler, supra note 67 (criticizing 2005 Survey), Bill Brenner, Security Blog
Log: Has CSI/FBI Survey Jumped the Shark?, SEARCHSECURITY.COM, July 21, 2006,
http://searchsecurity.techtarget.com/columnItem/0,294698,sid14_gci1202328,00.html;
Posting of Chris Walsh to Emergent Chaos, http://www.emergentchaos.com/archives/
2006/07/csifbi_survey_considered.html (July 16, 2006, 01:28) (challenging
methodology behind Survey).
71
See GORDON ET AL., supra note 66, at 1, 26 (listing Survey’s methodology); see
also Winkler, supra note 67 (suggesting that sample is not representative). For
example, the demographics have shifted in the 11 years of the survey. Id. In the 2005
Survey, the percentage of respondents who belong to the information sharing group
INFRAGARD and the percentage who use Intrusion detection system (IDS) software
differ widely from the results of an earlier version. Id.
72
See GORDON ET AL., supra note 66, at 19.
73
See Winkler, supra note 67.
74
Id.
75
See Tom Zeller, Jr., Breach Points Up Flaws in Privacy Laws, N.Y. TIMES, Feb. 24,
2005, at C1 (quoting Senator Dianne Feinstein: “Existing laws . . . are no longer
sufficient when thieves can steal data not just from a few victims at a time, but from
thousands of people with vast, digitized efficiency.”).
76
See, e.g., Press Release, U.S. Dep’t of Justice, Russian Hacker Sentenced to Four
Years in Prison for Supervising Criminal Enterprise Dedicated to Computer Hacking,
Fraud and Extortion and Victimizing Glen Rock Financial Services Company (July 25,
1346
University of California, Davis
[Vol. 41:1327
laptop theft is a common, low-tech way many criminals gather
personally identifiable information.77 Recently, ChoicePoint, a data
broker that has dealt with many data breach cases, concluded that “it
had focused so intently on preventing hackers from gaining access to
its computers through digital back doors that it had simply overlooked
real-world con artists strolling unnoticed through the front door.”78
Several studies have shown a recent decline in identity theft.79
District attorneys in the western United States have reported that a
majority of their identity theft arrestees are methamphetamine
addicts.80 Although some of these methamphetamine cases involve the
use of the Internet to facilitate identity theft,81 they also include nonSuperuser techniques like trash rifling, mail theft, or check washing.82
Another reason to doubt claims about the scope of online identity
theft is that many confuse mere data loss with identity theft.83 As one
journalist noted, “[W]hile high-profile data breaches are common,
2003), http://www.usdoj.gov/criminal/cybercrime/ivanovSent_NJ.htm (describing
hacker who broke into system and stole text file containing 3500 credit card
numbers).
77
See Erickson & Howard, supra note 57, at 2 (stating that “the growing number
of news stories about compromised personal records reveals a wide range of
organizational mismanagement and internal security breaches: lost hard drives and
backup tapes, employee theft, and other kinds of administrative errors”); Steve Lohr,
Surging Losses, But Few Victims, N.Y. TIMES, Sept. 27, 2006, at G1; David Stout, ‘Garden
Variety Burglary’ Suspected in Loss of Data, N.Y. TIMES, June 9, 2006, at A24
(describing theft of laptop containing information about 26.5 million military people).
78
Gary Rivlin, Keeping Your Enemies Close, N.Y. TIMES, Nov. 12, 2006, at 31
(finding that “in 2005 alone, more than forty phony businesses . . . had opened
accounts that gave them unfettered, round-the-clock access to the vital data
ChoicePoint maintains”); see also id. (summarizing conclusions of FTC report that
“criminal interlopers” who stole identities from ChoicePoint were “sloppy and
amateurish”).
79
See, e.g., John Leland, Identity Fraud Has Dropped Since 2003, Survey Shows, N.Y.
TIMES, Feb. 6, 2007, at A17 (reporting drop from 4.7% to 3.7% of Americans reporting
being victims of ID fraud in survey sponsored by banking industry).
80
See John Leland, Meth Users, Attuned to Detail, Add Another Habit: ID Theft, N.Y.
TIMES, July 11, 2006, at A1 (reporting 60 to 70% of identity theft cases in Denver and
100% in Spokane County, Washington are tied to methamphetamine users or dealers).
81
There are non-Superuser ways to use the Internet to assist identity theft. See
Robert Lemos, Google Queries Provide Stolen Credit Cards, CNET NEWS.COM, Aug. 19,
2004, http://news.com.com/2102-1029_3-5295661.html (describing technique
demonstrated at annual “Black Hat” conference for obtaining credit card numbers
using public Google searches).
82
See Leland, supra note 80. There is another way to interpret these anecdotes.
District attorneys may prosecute meth-addicted identity thieves more often because
they are easier to catch than the Superuser identity thieves. See id.
83
See Lohr, supra note 77.
2008]
The Myth of the Superuser
1347
there is no evidence of a surge in identity theft or financial fraud as a
result. In fact, there is scant evidence that identity theft and financial
fraud have increased at all.”84
Finally, consider briefly claims that terrorists are plotting to use
computer networks to threaten lives or economic well-being. There
has never been a death reported from an attack on a computer
network or system.85 In fact, despite claims to the contrary,86 many
doubt that an attack will ever successfully disable a significant part of
the Internet.87
Admittedly, too much reliance on anecdotes may smack of
hypocrisy. There is a risk of engaging in the “Myth of the Myth of the
Superuser.” There are limits to using opinions and qualitative
evidence to disprove the Myth because they share so much in common
with the anecdotes that fuel it. For this reason, I place greater stock in
the statistical observations made in the prior section and in Part IV.
3.
Obvious Overstatements of the Risk
Finally, some statements of the risk from Superusers are so
exaggerated that they are self-disproving. For example, Richard
Clarke, former Special Advisor to the President on Cybersecurity
under the Clinton and second Bush Administrations, often stated that
“digital Pearl Harbors are happening every day.”88 Even though the
phrase “digital Pearl Harbor” can refer to many different things —
attacks with the psychologically damaging effect, horrific loss of life,
terrifying surprise, size of invading force, or financial toll of the
84
Id.
See Brian Krebs, Feds Falling Short on Cybersecurity; Former Cybersecurity
Adviser Urges More Resources to Battle Cyberterror, WASHINGTONPOST.COM, Apr. 8,
2003, http://seclists.org/isn/2003/Apr/0036.html (quoting Richard Clarke testifying in
House Government Reform subcommittee: “For many, the cyber threat is hard to
understand; no one has died in a cyberattack, after all, there has never been a smoking
ruin for cameras to see.”).
86
See NAT. INFRASTRUCTURE ADVISORY COUNCIL, THE NATIONAL STRATEGY TO SECURE
CYBERSPACE 6 (2002), available at http://www.whitehouse.gov/pcipb/ [hereinafter
NATIONAL STRATEGY].
87
Frontline: Cyber War! (PBS television broadcast) (Mar. 20, 2003) (interview of
Scott Charney), available at http://www.pbs.org/wgbh/pages/frontline/shows/
cyberwar/interviews/charney.html. “[I]t’s not as easy to take down the Internet as
some might believe. There’s a lot of redundancy, a lot of resiliency in the system. . . . I
still think today the concern of a broad, sweeping global Internet attack that had long
enough staying power is not our number one threat today.” Id.
88
See Scott Berinato, The Future of Security, COMPUTERWORLD, Dec. 30, 2003,
available at http://www.computerworld.com/newsletter/0,4902,88646,00.html.
85
1348
University of California, Davis
[Vol. 41:1327
December 7, 1941 event — the claim is a horribly exaggerated
overstatement regardless of Clarke’s intended meaning.89
4.
The Scope of the Claim
These statistics and anecdotes suggest that many stories about
Superusers are exaggerated. Due to the incomplete nature of the
empirical evidence, this Article does not claim that the Superuser is
always a myth. Doubtless, some online harms are committed solely by
Superusers, and some Superusers probably cause significant harms.
Nevertheless, this survey of the empirical evidence should cast great
suspicion on the conventional account of unbridled power. And, as I
will demonstrate in the next Part, this empirical evidence should give
policymakers significant pause in light of the harms that flow from
attempting to regulate the Superuser.
II.
HARMS OF THE MYTH
Why should we care whether exaggerated arguments about
Superusers cause legislators to address risks that are unlikely to
materialize? Aside from dead-letter statutory prohibitions, are there
any other harms that flow from the Myth of the Superuser? The
answer is yes, there are significant harms. Moreover, the nearuniversal belief in the Myth means there has never been an accounting
of these harms, and thus we are doomed to repeat and extend them.
Below, I discuss five harms that flow directly from policies and laws
justified by the Myth.
A. Overbroad Laws
Congress typically responds to the Myth of the Superuser by passing
broad laws.
Generally, lawmakers broaden criminal and civil
prohibitions, giving law enforcement agencies sweeping new
authorities even though these can be used (and are used) against nonSuperusers.
In short, Congress overreacts. They fear an American version of
Onel de Guzman, the Philippines citizen who confessed to writing the
89
See Green, supra note 50. Clarke is not alone in his exaggeration. “Digital Pearl
Harbor” was not coined by him and it has been used by many over the past decade
and a half. Id. Other similar worry-phrases include “electronic Chernobyl,” “digital
Armageddon,” and “digital Waterloo.” See id.
2008]
The Myth of the Superuser
1349
“ILOVEYOU” Virus but escaped punishment because Philippines law
did not criminalize the type of harm he had caused.90
Furthermore, legislators tend to respond to the Myth by focusing on
statutory conduct elements rather than result, harm, intent, or
attendant circumstance elements.91 This makes sense. Conduct is
what makes the Superuser unusual, and the power they wield is often
For example, §
what some find offensive or threatening.92
1030(a)(5)(A)(i) of the CFAA prohibits “caus[ing] the transmission of
a program, information, code, or command.” This sweeping phrase
seems to encompass sending any data over the Internet. It applies to
all sorts of perfectly nonthreatening acts that ordinary users perform
every day.
Consider the following. In 2000, Bret McDanel worked for a
company that supplied email and voicemail accounts.93 He revealed
an internal security vulnerability to his employers, but they ignored
him.94 After leaving his job, McDanel sent an email message through
his former employer’s computer system in which he revealed the
vulnerability to 5600 of the company’s customers.95 The U.S. Attorney
prosecuted McDanel for violating § 1030(a)(5)(A).96
90
See Bryan Glick, Cyber Criminals Mock the Archaic Legal Boundaries,
COMPUTING, Jan. 4 2001, at 32.
91
In the Model Penal Code’s vocabulary, the elements of a criminal statute come
in four separate flavors:
conduct, results or harm, intent, and attendant
circumstances. See MODEL PENAL CODE § 1.13(9) (2001) (classifying conduct
elements into conduct, attendant circumstances, and results). For example, under 18
U.S.C. § 2252A(a)(5)(B) (2000) it is a crime to possess images of child pornography.
Parsing this prohibition into the four MPC categories:
[a]ny person who . . . knowingly [(intent)] possesses [(conduct)] any book,
magazine, periodical, film, videotape, computer disk, or any other material
that contains an image of child pornography [(results/harm)] that has been
mailed, or shipped or transported in interstate or foreign commerce by any
means, including by computer, or that was produced using materials that
have been mailed, or shipped or transported in interstate or foreign
commerce by any means, including by computer [(attendant
circumstances)].
18 U.S.C. § 2252A(a)(5)(B).
92
This bias applies even against those who use Superuser power benignly or for
morally or ethically good outcomes. See infra Part II.C.
93
See Chris Sprigman, The Federal Government’s Strange Cyber-Defamation Case
Against Bret McDanel: A Prosecution that Should Never Have Been Brought, FINDLAW,
Sep. 25, 2003, http://writ.news.findlaw.com/commentary/20030925_sprigman.html.
94
See id.
95
See id.
96
See id.
1350
University of California, Davis
[Vol. 41:1327
The prosecution’s only plausible theory must have been that
McDanel’s email messages amounted to “a program, information,
code, or command,” despite that the phrase usually applies to things
like computer viruses and worms.97 The U.S. Attorney’s interpretation
arguably met the plain text, because email messages are literally
“information.”98 Perhaps if Congress had been less fearful of the
Superuser, it could have drafted a more circumscribed, specific statute
that would not have applied to McDanel’s acts.99 Instead, McDanel
served one year and four months in prison.100 After he appealed, the
DOJ confessed error and dropped McDanel’s conviction.101
Like the overbroad language targeting conduct, § 1030(a)(5)(A)(ii)
prohibits
“access[ing]
[a
protected
computer]
without
authorization.”102 “Access” is not defined, and neither is “without
authorization.” Many courts have interpreted these vague terms
broadly.103
97
See id.
There is a second, perhaps more egregious, problem with such a theory. Only
those who commit “damage” are guilty of violating this provision. See 18 U.S.C. §
1030(a)(5)(A)(i) (2000 & Supp. V 2005). Damage is defined broadly, again perhaps
as a reaction to the Superuser, to include “any impairment to the integrity or
availability of data, a program, a system, or information.” Id. § 1030(e)(8). The
prosecution allegedly argued that the damage caused was McDanel’s employer’s
reputation. See Sprigman, supra note 93.
99
To be sure, it is imaginable, especially to one persuaded by the Myth of the
Superuser, that some Superuser criminal could use email messages to attack a system,
for example, as part of a Denial of Service attack. See generally JELENA MIRKOVIC ET AL.,
INTERNET DENIAL OF SERVICE: ATTACK AND DEFENSE MECHANISMS (2004) (discussing
Denial of Service attacks in depth). Even those who disagree with the prosecution of
McDanel might argue the conduct element should be broad to encompass this
hypothetical criminal. I contend that those making this argument have fallen prey to
the Myth of the Superuser. Even though this potential harm is covered by the broad
conduct element, that coverage comes at a cost of other harms from overbreadth, as
discussed below.
100
See Sprigman, supra note 93.
101
The government’s concession involved the damage element, not the conduct
element discussed here. See 18 U.S.C. § 1030(a)(5)(A)(i). In its motion to reverse the
conviction, the government conceded that “[o]n further review, in light of defendant’s
arguments on appeal, the government believes it was error to argue that defendant
intended an ‘impairment’ to the integrity of Tornado’s computer system.”
Government’s Motion for Reversal of Conviction at 3-4, United States v. McDanel, No.
03-50135 (9th Cir. Oct. 14, 2003), available at http://www.lessig.org/blog/
archives/govt.pdf.
102
18 U.S.C. § 1030(a)(5)(A)(ii).
103
Orin Kerr, Cybercrime’s Scope: Interpreting “Access” and “Authorization” in
Computer Misuse Statutes, 78 N.Y.U. L. REV. 1596, 1617 (2003) (noting “several recent
decisions point toward remarkably expansive interpretations of unauthorized access”).
98
2008]
The Myth of the Superuser
1351
Consider this example stemming from a civil lawsuit.104 A travel
agency had an employee write a computer program to collect, or
“scrape” its competitor’s prices from a website.105 The competitor
sued, alleging a violation of the “access to defraud” provision of the
CFAA, which relies on the aforementioned broad terms, “access” and
“authorization.”106 Although the defendants had merely accessed a
public website to copy publicly available information, the First Circuit
Court of Appeals held that they exceeded authorized access because
they arguably violated a confidentiality agreement between the
plaintiff and one of its former employees.107 In other words, this
broadly worded computer hacking and fraud statute proscribed mere
contract breach.
As a result of Congress’s expansion of these prohibitions, conduct is
no longer a meaningful, limiting principle for many federal computer
crimes. Because the Superuser’s conduct is hard to define, Congress
has given up trying to do so. So long as you merely “transmit” or
“access,” you have satisfied the conduct elements of the crime. These
elements have become low hurdles that, when cleared, place ordinary
users’ benign acts within the general reach of the prohibitions.108 In
McDanel’s case, the broad prohibitions meant sixteen months of
wrongful imprisonment.
The harm worsens with time. Each new hypothetical threat or
vague anecdote adds to the toolbox that Superusers are said to possess.
Lawmakers apply a ratchet to laws like § 1030, broadening substantive
provisions and increasing criminal penalties with nearly every
Congress.109 They are spurred on by law enforcement officials asking
104
The CFAA permits civil lawsuits for people harmed by the same prohibited acts
that are criminal under § 1030. 18 U.S.C. § 1030(g). Professor Kerr has noted the
spillover effect these civil lawsuits have on interpretations of criminal prohibitions, as
courts may be more inclined to entertain novel, aggressive theories of civil liability
that they might reject in the criminal context. See Kerr, supra note 103, at 1641-42
n.210.
105
EF Cultural Travel BV v. Explorica, Inc., 274 F.3d 577, 579-80 (1st Cir. 2001).
106
18 U.S.C. § 1030(a)(4); EF Cultural Travel BV, 274 F.3d at 581.
107
EF Cultural Travel BV, 274 F.3d at 582 (analyzing evidence in light of
preliminary injunction standard).
108
Other commentators have written about this feature of computer crime law, but
they are split about whether it is desirable. Compare Kerr, supra note 103, at 1647-48
(arguing for expansive interpretation of “access” within § 1030, placing greater weight
on meaning of “authorization”), with Patricia L. Bellia, Defending Cyberproperty, 79
N.Y.U. L. REV. 2164, 2254 (2004) (advocating for narrower meaning for access as “the
more natural” reading).
109
The Senate Report on the 1996 amendments to the Act states:
1352
University of California, Davis
[Vol. 41:1327
for help tackling new threats on the horizon.110 Meanwhile, onceinnocent behavior begins to fall into new classes of prohibited
conduct.
Broad statutes are not only a problem when they lead to the
conviction of the inculpable. They also raise civil liberties concerns by
enabling what I call “investigatory overbreadth.” Broad conduct
elements lead to enormous suspect pools. Imagine there has been an
attack on a corporate web server. Because the court in the travel
agency website dispute111 construed “without authorization” to apply
to those who merely breach contractual duties, the acts of all
employees and contractors must be scrutinized.112 Because the
McDanel court held that email messages to third parties constituted
the “transmission of information,” the private email messages or
instant messages of customers and other outsiders should also be
scrutinized.113
Thus, investigatory overbreadth refers to how broad conduct
elements place no limit on the number or type of people who are
suspects.114 Compounding the problem, broad conduct elements
make it easier for police to establish probable cause to search the
belongings of suspects. This is further exacerbated by the fact that
most Internet surveillance laws do not require notice to the party
As computers continue to proliferate in businesses and homes, and new
forms of computer crime emerge, Congress must remain vigilant to ensure
that the Computer Fraud and Abuse Act statute is up-to-date and provides
law enforcement with the necessary legal framework to fight computer
crime.
S. REP. NO. 104-357, pt. 2, at 5 (1996); see also United States v. Middleton, 231 F.3d
1207, 1212 (9th Cir. 2000) (stating “Congress has consciously broadened [§ 1030]
since its original enactment”).
110
See supra note 18 (listing major revisions to § 1030).
111
See supra notes 102-103 and accompanying text.
112
See supra note 105.
113
See supra note 93.
114
A related trend is the creation of laws that punish the possession or use of a
particular technology, rather than the harm caused by the technology. See Joseph M.
Olivenbaum, Ctrl-Alt-Delete: Rethinking Federal Computer Crime Legislation, 27 SETON
HALL L. REV. 574, 575-76 (1997) (explaining that “[t]o the extent that [computer
crime laws] focus on technological means, rather than on the harm caused by a
defendant’s conduct, those statutes tend towards overbreadth by sweeping within their
ambit anyone who uses the means regardless of result”); Douglas Thomas, Criminality
on the Electronic Frontier: Corporality and the Judicial Construction of the Hacker, in
CYBERCRIME: LAW ENFORCEMENT, SECURITY, AND SURVEILLANCE IN THE INFORMATION AGE
18 (Douglas Thomas & Brian D. Loader eds., 2000).
2008]
The Myth of the Superuser
1353
surveilled.115 Thus, law enforcement officials could read the email
messages and obtain the browsing habits of dozens or more people
without jeopardizing the investigation. The combination of broad
prohibitions and low surveillance standards gives the police both the
incentive and the means to cast out larger and more invasive dragnets.
Suppose that in our web server hypothetical, Congress had clarified
that certain crimes could be conducted only by outsiders.116 In this
situation, the pool of suspects would have been significantly smaller.
Likewise, this would be the case if Congress had clarified that
“transmitting information” did not apply to mere email
communications, contra McDanel.117 Congress, however, is loath to
narrow conduct elements, not because it is convinced that insiders or
people like McDanel deserve punishment, but more likely because it
worries that a Superuser’s acts will slip outside a narrow prohibition.
But is the possibility of a highly unlikely criminal evading conviction
through a loophole worse than routinely investigating and prosecuting
people like McDanel for seemingly innocent acts?
B. Unduly Invasive Search and Seizure
Part of what is terrifying about the Superuser is how the Internet
allows him to act anonymously. He can hop from host to host and
country to country with impunity. To find the Superuser, the police
need better search and surveillance authorities, better tools, and the
latitude to pursue creative solutions for piercing anonymity.
But broad search authorities can be used unjustifiably to intrude
upon civil liberties. Search warrants for computers are a prime
example, because the judges who sign and review these warrants
115
See 18 U.S.C. § 2511(2)(a)(ii)(B) (2000) (forbidding providers from
“disclos[ing] the existence of any interception or surveillance” conducted pursuant to
court order under Wiretap Act and providing civil damages for failing to comply); id.
§ 2703(c)(3) (2000 & Supp. V 2005) (dispensing with notice requirement under
Stored Communications Act for government access, with appropriate process, to
“records or information” about subscriber); id. § 3123(d)(2) (2000 & Supp. V 2005)
(similar provision for pen registers and trap and trace devices). But see id. §
2703(b)(1)(B) (2000) (requiring “prior notice” for access to certain types of
subscriber content information); id. § 2705 (2000) (providing mechanism for delaying
notice required by § 2703).
116
Cf. id. § 1030(a)(3) (2000) (defining criminal attacks on government systems to
exclude certain insiders); id. § 1030(a)(5)(A)(i) (2000 & Supp. V 2005) (defining
nonaccess attacks on protected computers to apply only to one “without
authorization” but omitting, by implication, insiders who act “in excess of
authorization”).
117
See supra note 93.
1354
University of California, Davis
[Vol. 41:1327
usually authorize sweeping and highly invasive searches justified by
storytelling about a particular species of Superuser we might call the
“Data Hider.”
Agents seeking computer search warrants consider it standard
practice to tell stories in supporting affidavits about the sophisticated
technology that can be used to hide data.118 According to this
boilerplate, criminals are known to use steganography,119 kill
switches,120 and encryption to hide evidence of their crimes.121 These
agents also assert that file names and extensions are almost
meaningless, because users can easily change this information to hide
data.122
These assertions are important, because courts have
repeatedly held that each file in a computer is a separate “container”
over which a person has a reasonable expectation of privacy under the
Fourth Amendment and for which the police must usually establish
independent probable cause to open.123
Convinced of the prowess of the Data Hider, a typical judge will
usually sign a warrant that authorizes (1) the search of every single
file on subject computers, (2) the search of hard drive parts that do
not even store files,124 and (3) offsite searches, where data is
118
See U.S. DEP’T OF JUSTICE, COMPUTER CRIME & INTELLECTUAL PROP. SECTION,
SEARCHING AND SEIZING COMPUTERS AND OBTAINING ELECTRONIC EVIDENCE IN CRIMINAL
INVESTIGATIONS (2002), available at http://www.usdoj.gov/criminal/cybercrime/
s&smanual2002.htm [hereinafter DOJ MANUAL] (offering model search warrant that
includes language to justify offsite search: “Searching computer systems requires the
use of precise, scientific procedures which are designed to maintain the integrity of
the evidence and recover ‘hidden,’ erased, compressed, encrypted or passwordprotected data.”).
119
Steganography is defined and discussed infra in Part IV.B.1.d.
120
Kill switches are commands or hardware devices that can be triggered to cause
the deletion of data. See Timothy Roberts, Protecting Against Digital Data Thefts with a
‘Kill Switch,’ SAN JOSE BUS. J., Dec. 30, 2005, available at http://sanjose.bizjournals.com/
sanjose/stories/2006/01/02/story5.html.
121
See, e.g., United States v. Comprehensive Drug Testing, Inc., 473 F.3d 915, 961
(9th Cir. 2006) (assessing search warrant affidavit that spun tales about users
mislabeling files, using encryption, and using steganography).
122
File extensions are the parts of file names that, by convention, reveal the broad
type of data stored within. On Windows computers, these extensions are usually the last
three letters of the file name, following the final dot.
For example,
“SuperuserArticle.doc” has an extension of “.doc.” The “.doc” extension signifies that
the file is a Microsoft Word document. But any user can change the filename to mask
the extension, making the Word document, for example, appear to be an MP3 music file.
123
See United States v. Carey, 172 F.3d 1268, 1273 (10th Cir. 1999); Kerr, supra
note 45, at 554-57.
124
These areas of “latent data” are often unknown to most computer users. Some
examples include the swap file, deleted space, file slack, and RAM slack. See Kerr,
2008]
The Myth of the Superuser
1355
forensically examined for months or maybe even years.125 In
upholding the scope of these searches, reviewing courts make bare and
broad proclamations about what criminals do to hide evidence.126
These broad pronouncements are built upon nothing more than the
agent’s assertions and the judge’s intuitions about computer
technology.
For example, the Ninth Circuit Court of Appeals has held that
“[c]omputer records are extremely susceptible to tampering, hiding,
or destruction, whether deliberate or inadvertent.”127 To support this
claim, the Ninth Circuit quoted a district court opinion from
Vermont,128 which in turn cited a district court opinion from
Florida.129 The Florida court based its conclusion about data
tampering on what an agent said he had been told by a Customs
Service forensic computer expert.130 Such is the path from the Myth of
the Superuser to binding court of appeals case law.
In reality, if criminals tend not to hide data inside obscured file
names or unusual directories, judges might feel compelled to ask the
police to cordon off parts of a computer’s hard drive.131 The law
supra note 45, at 542 (stating that “[c]omputers are also remarkable for storing a
tremendous amount of information that most users do not know about and cannot
control”).
125
See United States v. Hill, 459 F.3d 966, 978 (9th Cir. 2006) (upholding search
warrant that allowed blanket search through all files on hard drive); United States v.
Adjani, 452 F.3d 1140, 1150 (9th Cir. 2006) (same).
126
See Hill, 459 F.3d at 978 (finding that “[c]riminals will do all they can to
conceal contraband, including the simple expedient of changing the names and
extensions of files to disguise their content from the casual observer”); Adjani, 452
F.3d at 1150 (stating that “[c]omputer files are easy to disguise or rename, and were
we to limit the warrant to such a specific search protocol, much evidence could escape
discovery simply because of Adjani’s (or Reinhold’s) labeling of the files documenting
Adjani’s criminal activity . . . [t]he government should not be required to trust the
suspect’s self-labeling when executing a warrant”).
127
Hill, 459 F.3d at 978.
128
United States v. Hunter, 13 F. Supp. 2d 574, 583 (D. Vt. 1998).
129
United States v. Abbell, 963 F. Supp. 1178, 1199 (S.D. Fla. 1997).
130
Id.
131
In a dissenting opinion, Judge Fisher of the Ninth Circuit, who wrote Adjani,
seemed to argue for requiring this kind of additional proof, albeit in an atypical
context, that is, in a case involving the criminal investigation arising from the Major
League Baseball “BALCO” steroid scandal. United States v. Comprehensive Drug
Testing, Inc., 473 F.3d 915, 919-20 (9th Cir. 2006). Judge Fisher opined that the
government had “made misleading statements in . . . search warrant applications,” in
part because “[t]he government did not have any evidence or reason to believe that
CDT had engaged in steganography, boobytrapping computers, or any type of data
destruction or alteration.” Id. at 961.
1356
University of California, Davis
[Vol. 41:1327
enforcement community fears such a result.132 It prefers instead to use
computer forensics tools that treat the hard drive as a unitary pool of
data through which to search.133
So where does this particular myth end and reality begin? Common
sense suggests that some criminals are paranoid enough to hide
evidence. But it is also highly improbable that all criminals are likely
to use these tactics. Home computer users committing relatively
nontechnological crimes — death threats or extortion via email, for
example — may have less incentive to hide evidence and no access to
the tools required to do so. Painting all criminals in every warrant
application as uniformly capable of hiding information is a classic
example of the Myth.
In accepting the bare assertion that every computer user is a
potential Data Hider, judges may fail to uphold the Fourth
Amendment rights of those searched. In some cases, constraints on
the allowable scope of the search of a hard drive may be sensible and
even constitutionally mandated.134
132
See DOJ MANUAL, supra note 118, § II.C (stating, “[f]or example, it is generally
unwise to limit a search strategy solely to keyword searches”); Kerr, supra note 45, at
576 (explaining “[t]he computer forensics process calls for ex-post standards, not exante rules”).
133
See Kerr, supra note 45, at 538. This is not to say that computer forensics could
not be executed in a more limited, privacy-sensitive manner. If a court signed a
warrant that required the police to avoid particular parts of a hard drive, forensics
experts would be able to use most of their tools to do this kind of analysis. See Brief of
Amici Curiae Computer Forensics Researchers and Scientists in Support of Appellant
and Reversal of the Denial of the Motion to Suppress at 22, United States v. Andrus,
No. 06-3094 (10th Cir. June 1, 2007), 2007 WL 3264595, at *22 (brief co-authored by
this Article’s author) (“Furthermore, [the commonly used computer forensics tool]
EnCase makes it possible to create a filter to exclude specified parts of the hard drive
from review: a forensic technician merely needs to click on specific folders from
among all of the folders on the hard disk to include them and only them in an
operation.”).
134
For example, in a search of a hard drive for evidence of music illegally
distributed over peer-to-peer networks, it may make sense to limit the search to the
computer directories used by that particular type of peer-to-peer software. Just as a
warrant to search for a gun cannot be used to support a search through stacks of paper
on a desk, cf. United States v. Ross, 456 U.S. 798, 820 (1992) (“A lawful search of
fixed premises generally extends to the entire area in which the object of the search
may be found . . . .”), agents should not be allowed to look for music where it cannot
be found. Compare United States v. Walser, 275 F.3d 981, 986 (10th Cir. 2001)
(upholding scope of computer search because agent “searched for relevant records in
places where such records might logically be found. . . . [He] selectively proceeded to
the ‘Microsoft Works’ sub-folder on the premise that because Works is a spreadsheet
program, that folder would be most likely to contain records relating to the business
of drug trafficking”), with United States v. Carey, 172 F.3d 1268, 1274 (10th Cir.
2008]
The Myth of the Superuser
1357
Allowing computer-wide searches in every case dramatically
impinges on privacy. As hard drive capacity grows, the incentive for
computer users to delete old files diminishes.135 Today’s computers
can contain tens of thousands of letters, email messages, business
records, and financial documents, stretching back years.136 In the
computer context, courts have interpreted the plain view rule to mean
that evidence of any crime found during a computer search can be
used to prosecute the computer’s owner, even if it is unrelated to the
crime recited in the warrant.137 Commentators have likened hard
drive-wide searches to the general warrants that incensed the
Founding Fathers.138 Even if this is not a perfect comparison, these
searches are the closest thing to general warrants that we have in
modern police practice. By succumbing to the Myth, judges have
given the police the power to search at odds with Fourth Amendment
protections.
C. Guilt by Association
Another harm results when policymakers confuse power and evil.
This mistake is borne of a flawed syllogism: Power can be used online
to cause harm, Superusers are powerful, and therefore, Superusers are
harmful. This ignores the fact that many Superusers cause no harm
and may even cause great benefit. As a result of this flawed view,
benign or beneficial Superusers are branded illicit, and in the extreme
case, they are sued or prosecuted for doing nothing except wielding
1999) (suppressing evidence found on computer because after finding one image of
child pornography, “[w]hen he opened the subsequent [similarly named] files, he
knew he was not going to find items related to drug activity as specified in the
warrant”). Obviously, if law enforcement agents have any particularized reason to
suspect that this music distributor is likely to obscure data, the affidavit should reflect
this fact, and the warrant should allow more scrutiny of the hard drive.
135
Cf. Paul Festa, Google to Offer Gigabyte of Free E-mail, CNET NEWS.COM, Apr. 1,
2004,
http://www.news.com/Google-to-offer-gigabyte-of-free-e-mail/2100-1032_35182805.html?tag=item (explaining that “Google will offer enough storage so that the
average email account holder will never have to delete messages”).
136
See Kerr, supra note 45, at 541-42 (“Computer hard drives sold in 2005
generally have storage capacities . . . roughly equivalent to . . . the amount of
information contained in the books on one floor of a typical academic library.”).
137
See id. at 576-77.
138
See id. at 566 (“Narrowing or even eliminating the plain view exception may
eventually be needed to ensure that warrants to search computers do not become the
functional equivalent of general warrants.”).
1358
University of California, Davis
[Vol. 41:1327
their power.139 This is guilt by association of an especially pernicious
and illogical form.
Felten, a professor of Computer Science and Public Affairs at
Princeton University, has suffered from this kind of overreaction.
Felten’s research focuses on DRM and computer security, using
especially applied methods to actively try to circumvent software
security to expose flaws.140 Under threat of a lawsuit, Felten was once
forced to delay presenting his research.141 He now consults regularly
with lawyers before undertaking sensitive projects and consuming
time and energy better spent on research.142
D. Wasted Investigatory Resources
Because Superusers can evade detection and identification, they are
difficult to find. In my opinion, while the DOJ does a very good job of
capturing and punishing dim hackers, the smart ones tend to get
away. Given enough money, time, and tools, law enforcement
agencies could catch some Superusers, but with the same resources,
they could find many more non-Superusers instead.
Even though law enforcement tends primarily to capture nonSuperuser criminals, the DOJ raises the specter of the Superuser
criminal whenever it discusses computer crime with Congress.143
139
See supra note 114.
See, e.g., Posting of Ed Felten to Freedom to Tinker Blog, http://www.freedomto-tinker.com/?p=975 (Feb. 14, 2006, 09:19) (summarizing DRM research on Sony
rootkit).
141
See John Markoff, Scientists Drop Plan to Present Music-Copying Study That
Record Industry Opposed, N.Y. TIMES, Apr. 27, 2001, at C5.
142
Letter from Edward W. Felten & J. Alex Halderman, Princeton Univ., to Office
of the Gen. Counsel, U.S. Copyright Office 7 (Dec. 1, 2005), available at
http://www.freedom-to-tinker.com/doc/2005/dmcacomment.pdf (“Researchers like
Professor Edward Felten and Alex Halderman waste valuable research time consulting
attorneys due to concerns about liability under the DMCA.”) (responding to
Exemption to Prohibition on Circumvention of Copyright Protection Systems for
Access Control Technologies, 71 Fed. Reg. 68,472 (Nov. 27, 2006) (to be codified at
37 C.F.R. pt. 201)).
143
See, e.g., The Cyber Security Enhancement Act of 2001: Hearing on H.R. 3482
Before the Subcomm. on Crime of the H. Comm. on the Judiciary, 107th Cong. (2002)
(statement of John G. Malcolm, Deputy Assistant Att’y Gen. of the United States),
available at http://www.cybercrime.gov/HR3482_01Testimony.htm (describing
Russian hackers and citing CSI/FBI statistics); Department of Justice’s Efforts to Fight
Cybercrime: Hearing Before the Subcomm. on Crime of the H. Comm. on the Judiciary,
107th Cong. (2001) (citing statement of Michael Chertoff, Assistant Att’y Gen. of the
United
States),
available
at
http://www.cybercrime.gov/cybercrime61201_
MChertoff.htm (describing hackers from Russia and Eastern Europe, cyberterrorists,
140
2008]
The Myth of the Superuser
1359
Congress often responds by increasing the resources and tools at the
DOJ’s disposal, focusing on the hardest cases, which probably
represent a small fraction of the victims and harm.144
E. The Myth and Legal Scholarship
Scholars too often fall prey to the Myth.145 By doing this, at the very
least, they shift attention away from the proper heart of most online
debates — the problems posed by ordinary users. Worse, the
prescriptions arising from these Myth-based arguments are often
fundamentally flawed.
Consider again the DRM debate. Scholarly critics of DRM point to a
paper written by four Microsoft engineers, entitled The Darknet and the
Future of Content Distribution, as proof of the ineffectiveness of laws
like the Digital Millennium Copyright Act (“DMCA”).146 The first
premise of the Darknet paper is the sentence most often misused:
“Any widely distributed object will be available to a fraction of users in
a form that permits copying.”147 In other words, in the battle between
lock builders and lock pickers, the Darknet authors assume that the
lock-picking Superusers have the upper hand.
Legal scholars who cite this proposition often miss the fact that it is
merely an assumption.148 The Darknet paper authors do not prove the
assumption with rigor, but instead take it as a starting point. Others,
attacks on critical infrastructures, and dubious statistics).
144
See supra Part I.A.2 (describing Congress’s approach to modifying CFAA).
145
See infra notes 146-59 (discussing legal scholars’ use of so-called “darknet”
hypothesis). Student note authors seem to fall prey to the Myth more often than their
counterparts in the professorial ranks. See, e.g., Stephen W. Tountas, Note, Carnivore:
Is the Regulation of Wireless Technology a Legally Viable Option to Curtail the Growth of
Cybercrime?, 11 WASH. U. J.L. & POL’Y 351, 376 (2003) (“Given the devastation of
September 11, along with sophisticated tactics such as steganography, it is in
Congress’[s] best interest to disregard Carnivore’s constitutional issues” (footnote
omitted)). It may be that student authors are more careless or prone to logical
missteps in their analyses. On the other hand, it may be that student authors are more
aware of advanced technology and more willing to consider the implications of the use
of advanced technology.
146
See Peter Biddle et al., The Darknet and the Future of Content Distribution (2002),
available at http://crypto.stanford.edu/DRM2002/darknet5.doc. The scholar most
associated with bringing the Darknet paper into mainstream legal scholarship is Fred
von Lohmann. Fred von Lohmann, Measuring the Digital Millennium Copyright Act
Against the Darknet: Implications for the Regulation of Technological Protection
Measures, 24 LOY. L.A. ENT. L. REV. 635, 641 (2004).
147
Biddle et al., supra note 146, at 2.
148
Id.
1360
University of California, Davis
[Vol. 41:1327
however, have cited the first premise of the Darknet paper as proven
fact,149 which it is not.150
Compounding the problem, those who cite the first premise tend to
misread exactly what it says. The first premise is a statement about
possibilities, not inevitabilities.151 The authors do not (nor could they)
contend that the world’s Superusers have the skill, time, and interest
to crack every single piece of DRM-protected content. Theirs is a
more modest point about incentives, studded with caveats: only
works that are “widely distributed” satisfy the claim; only a “fraction
of users” can copy these works; even vulnerable works that “permit”
copying will not necessarily be copied.152
The Darknet paper, in sum, supports several propositions at odds
with the conventional scholarly perception of the paper. First, DRMprotected copies of unpopular music may never be broken, because no
Superusers will unlock them. Likewise, because of what I call the
“limits of human bandwidth,”153 — the idea that Superusers break
DRM at a fixed rate — even protection schemes used for popular
music may have to wait in line for an interested Superuser to come
along.154 For some content owners, DRM systems that remain
unbroken for months or years are good enough.155 This is especially
true because users can always be convinced or forced to upgrade to
149
See Neil Weinstock Netanel, Impose a Noncommercial Use Levy to Allow Free
Peer-to-Peer File Sharing, 17 HARV. J.L. & TECH. 1, 9-10 (2003); von Lohmann, supra
note 146, at 640.
150
The reason the Darknet paper authors felt no need to prove the first premise is
because their aim was to comment on what happens after Superusers have acted.
Their central argument was that small, informal, closed-but-interconnected networks
can efficiently distribute libraries of copyrighted works that “approach the aggregate
libraries that are provided by the global darknets [such as the peer-to-peer networks]
of today.” Biddle et al., supra note 146, at 9. Yesterday’s tight-knit circles of cassetteswapping teenagers have been replaced by larger groups with fast Internet
connections, complex software, powerful computers, and giant hard drives. So long
as some Superusers feed these darknets (again, this is just an assumption), these
darknets will thrive and be very difficult to detect and shut down. People who cite the
Darknet paper often mistake the starting point for the conclusion. See Netanel, supra
note 149, at 9-10; von Lohmann, supra note 146, at 640.
151
See Biddle et al., supra note 146, at 2.
152
Id.
153
See infra note 241 and accompanying text (discussing limits of human
bandwidth).
154
See Nate Anderson, Hacking Digital Rights Management, ARS TECHNICA, July 18,
2006, http://arstechnica.com/articles/culture/drmhacks.ars (noting that Microsoft’s
DRM system for audio “has not been widely breached” since late 2001).
155
See id. (noting that even imperfect DRM schemes “may be good enough for
most [record] labels”).
2008]
The Myth of the Superuser
1361
newer, stronger versions of DRM.156 In fact, despite how it is
portrayed and regarded in the scholarly community, the Darknet paper
is surprisingly optimistic about certain aspects and types of DRM. For
example, the authors note that “[e]xisting DRM-systems typically
provide protection for months to years.”157 In other words, DRM is
often good for a few months’ head start.
To be sure, the Darknet paper casts serious doubts on the ability of
DRM to stop all copyright infringement. The authors try to temper
expectations that laws like the DMCA will be a “silver bullet” against
the spread of unauthorized copies of copyrighted works.158 Thus, in
the debate over DRM, the paper stands squarely on the side of those
who doubt DRM’s status as a panacea.
Nevertheless, the paper’s conclusions have been overstated by
scholars tempted by the Myth. The sound-bite version of the paper’s
conclusion is the claim that powerful users will circumvent any
protection scheme released.159 This sound bite is intuitive, appealing,
and wrong.160
156
See Randal C. Picker, Rewinding Sony: The Evolving Product, Phoning Home and
the Duty of Ongoing Design, 55 CASE W. RES. L. REV. 749, 766-68 (2005); Jonathan L.
Zittrain, The Generative Internet, 119 HARV. L. REV. 1974, 2019 (2006).
157
In the conclusion to the paper, the authors go so far as to say that if darknets
tend to be isolated from one another (a possibility the authors implicitly seem to
doubt), then some particularly weak classes of DRM “are highly effective.” Biddle et
al., supra note 146, at 15.
158
However, the Darknet paper authors are optimistic about the law’s ability to
disrupt aspects of digital copyright infringement. In fact, the paper appears to have
been written in response to the success of lawsuits against centrally run services such
as Napster, Gnutella, and Kazaa. See id. at 5-8. The paper seems decidedly pessimistic
about the ability of such centralized services to resist legal challenges for long. See id.
at 7-8. Even the DMCA is called a “far-reaching (although not fully tested) example
of a law that is potentially quite powerful.” Id.
159
See Netanel, supra note 149, at 9 (citing Darknet paper, along with another
source, for proposition that, “[i]n fact, computer security experts maintain that no
technological barrier can ultimately prevail over determined hackers who have
physical access to the encrypted items”).
160
The Darknet paper is a good model of a measured, careful way of dealing with
the Superuser. In its last section, the paper calls for further empirical work about the
nature of the darknets, consistent with my recommendation for a more searching
empirical inquiry to back up Superuser claims in Part IV. See Biddle et al., supra note
146, at 15 (stating that “[i]t appears that quantitative studies of the effective ‘diffusion
constant’ of different kinds of darknets would be highly useful in elucidating the
dynamics of DRM and the darknet”).
1362
University of California, Davis
[Vol. 41:1327
III. MYTH ORIGINS AND PERPETUATION
Why is it so difficult to assess Superuser risks — to measure the
power and reach Superusers command? There are three parts to the
answer, which together describe a pervasive fear-processing machine.
The machine operates on fear, and in particular, a documented,
widely held fear of the Internet.161 We fear the Internet for both
rational and irrational reasons; our fears are created or magnified by
the media. Just-so stories about imaginable (but implausible) new
powers that evil Superusers may wield are fed into the machine, which
in turn spits out miscalculated risks, misguided advocacy, poor
policymaking, and the litany of harms outlined in Part II.
So far, this is a fairly conventional account, as the basic connection
between fear and risk assessment is well-known in the social science
literature. Specifically, the literature surveyed in Part III.A identifies
heuristics and biases which lead us to miscalculate risk when faced
with fear. These heuristics and biases are the valves, gears, and
bearings of the fear-processing machine. Although this conventional
account can explain, in part, the Myth of the Superuser, I extend and
challenge the literature by focusing on two much more idiosyncratic
features of the Internet’s fear-processing machine.
The first variation from the conventional account stems from the
malleability of software. Software is easy to change in fundamental
ways, but those who fall prey to the Myth of the Superuser
misinterpret or misuse this fact. As I discuss in Part III.B, although
software is malleable, many observers have confused the possible with
the inevitable, partly because they have been confused by Professor
Lessig’s thoughts about code.
Misunderstandings about the
malleability of software feed new stories of fear and risk to the
machine at a rate rarely seen with real world stories.
Second, I focus on experts, the machine’s operators. In other
situations, experts try to temper public fear to slow the machine’s
output to a manageable rate. With online risk, the machine is
unmanned. As Part III.C explains, even experts cannot assign
probabilities to online risks.
Taken together, these features — fear, code malleability, and expert
abandonment — turn the conventional account into a much more
troubling one. It becomes more difficult to remedy and a more
pressing a concern than analogous problems offline. To begin the
assessment, I start with fear.
161
See infra Part III.A.1.b.
2008]
The Myth of the Superuser
1363
A. Myth Origins: Fear
1.
Fear
During the past thirty-five years, sociologists and psychologists have
developed an extensive literature about fear, focusing in particular on
human responses to fear and the effect of those responses on decisionmaking.162 Scholars have recently imported this literature into
theories of law, particularly with respect to environmental risks, such
as global warming.163 To date, no one has applied this fear and risk
literature to Internet regulation. This approach holds great promise,
in light of both the widely held and well-documented fear of the
Internet as well as the pre-eminence of rhetoric and storytelling in
policy debates about online regulation.
a.
The Effects of Fear
Fear causes laypeople and policymakers to exaggerate some risks
and to downplay others in spite of the actual probability that those
risks will occur. In Laws of Fear, Cass Sunstein surveys psychological
and sociological explanations.164
Sunstein focuses on two as
particularly important: the availability heuristic and probability
neglect.165
The availability heuristic is the psychological tendency to give
greater weight to scenarios one can imagine.166 The heuristic is
triggered primarily with risks that are both familiar and salient.167
Familiar risks are those we can readily imagine.168 Because the risks
from smoking are more familiar than the risks from sunbathing,
Sunstein notes, people are more likely to exaggerate the risk from
162
See generally HEURISTICS AND BIASES: THE PSYCHOLOGY OF INTUITIVE JUDGMENT
(Thomas Gilovic, Dale Griffin, & Daniel Kahneman eds., 2002) (collecting important
papers in heuristics and biases scholarship); JUDGMENT UNDER UNCERTAINTY:
HEURISTICS AND BIASES 3, 11-14 (Daniel Kahneman, Paul Slovic, & Amos Tversky eds.,
1982) (same); PAUL SLOVIC, THE PERCEPTION OF RISK (2000) (examining disconnect
between expert measurements of risk and public perception of risk).
163
See CASS SUNSTEIN, LAWS OF FEAR: BEYOND THE PRECAUTIONARY PRINCIPLE 36
(2005).
164
See id. at 35-49.
165
See id.
166
See id. at 36-39.
167
See id. at 37.
168
See id.
1364
University of California, Davis
[Vol. 41:1327
smoking.169 Salient risks trigger a similar reaction. Viewing a house
fire, for example, is likely to have more impact than reading about it in
the newspaper because of the saliency of witnessing the fire
firsthand.170
Probability neglect is the tendency to pay little attention to the
probability of a risk occurring, often experienced with worst case or
emotionally charged risks.171 The risk of death from airplane crashes,
for example, tends to trigger stronger reactions than the other, more
likely causes of death because of probability neglect.172 Sunstein notes
that “when intense emotions are engaged, people tend to focus on the
adverse outcome, not on its likelihood.”173
b.
Fear of the Internet
The widely held fear of the Internet connects these observations to
the Myth of the Superuser. Prior literature demonstrates a strong
connection between fear and new technology.174 In perhaps the most
comprehensive summary of studies about information technophobia,
Professor Mark Brosnan reports large segments of the population
suffer from some fear of computer technology.175
Repeated surveys spotlight people’s fear of the Internet. A survey
commissioned by the U.K. government found that “[f]ear of [I]nternet
crime is now more prevalent than concerns about more conventional
crimes such as burglary, mugging and car theft.”176 In a 2003 survey,
169
See id.
See id. (citing Amos Tversky & Daniel Kahneman, Judgment Under Uncertainty:
Heuristics and Biases, in JUDGMENT UNDER UNCERTAINTY, supra note 162, at 3, 11-14).
171
See id. at 39-41.
172
See id. at 39-40. Sunstein notes that the availability heuristic also plays a role in
this particular example. See id. at 40. The availability heuristic will cause a person to
think airplane crashes are more likely to occur than their actual incidence, while
probability neglect will cause people to ignore the probability of an airplane crash
completely. See id. at 39.
173
Id. at 64; see also id. at 40 (explaining that “vivid images of disaster . . . crowd[]
out probability judgments”).
174
See, e.g., MARK J. BROSNAN, TECHNOPHOBIA: THE PSYCHOLOGICAL IMPACT OF
INFORMATION TECHNOLOGY 10-36 (1998) (summarizing studies about extent of fear of
information technology).
175
Id. at 12. Brosnan cites studies, for example, finding 50% of college students
registering as technophobic. Id. (citing Larry D. Rosen & Phyllisann Maguire, Myths
and Realities of Computerphobia: A Meta-Analysis, 3 ANXIETY RESEARCH 175 (1990)).
176
See Helen Carter, Internet Crime Eclipses Burglary in Survey of Perceived Risks,
THE GUARDIAN, Oct. 9, 2006, http://www.guardian.co.uk/technology/2006/oct/09/
news.crime.
170
2008]
The Myth of the Superuser
1365
the Pew Internet and American Life Project discovered that 49% of
Americans fear that terrorists might cripple American utilities, banks,
or corporations through cyberattacks.177 Two years earlier, the same
organization discovered that over 70% of Americans were “concerned”
or “very concerned” about child pornography, credit card theft,
hackers, and organized terrorism online.178
We fear the Internet on several levels. First, we fear that the world
is becoming less comprehensible to the average person. We fear that
increasing technological complexity masks a terrifying fragility: the
world seems one cascading failure away from becoming unplugged,
taking away all of the essential services we have migrated online in the
past decade.179
Second, we fear malicious Superusers on the Internet for several
reasons. We imagine the Internet teeming with all kinds of evildoers,
from simple predators to “Supercriminal” Superusers, such as
organized crime figures, terrorists, and war fighters.180 Worse, they
have tools unlike any seen before; even everyday Internet applications
like email and the web are used to perpetrate frauds and harm
children with terrifying efficiency.181
These tools are nothing
compared to the more powerful and more inscrutable ones we fear
they wield.182 While it would be hard to claim that the fear of
177
MEMORANDUM FROM LEE RAINIE, PEW INTERNET & AM. LIFE PROJECT, SURVEY WITH
FEDERAL COMPUTER WEEK MAGAZINE ABOUT EMERGENCIES AND THE INTERNET 1 (Aug. 31,
2003), available at http://www.pewinternet.org/pdfs/PIP_Preparedness_Net_Memo.pdf.
178
PEW INTERNET & AM. LIFE PROJECT, FEAR OF ONLINE CRIME: AMERICANS SUPPORT
FBI INTERCEPTION OF CRIMINAL SUSPECTS’ EMAIL AND NEW LAWS TO PROTECT ONLINE
PRIVACY 9 (2001), available at http://www.pewinternet.org/pdfs/pip_fear_of_crime.pdf
(showing over 70% of Americans “concerned” or “very concerned” about child
pornography, credit card theft, hackers, and organized terrorism online).
179
See TAYLOR, supra note 29, at xiii (“Conversely, fear of computer technology
complements our perennial cultural concern that we cannot ultimately control our
technological curiosity.”). Paul Taylor links the fear of computer technology to “the
historical range of cultural expressions that give [the fear] voice,” citing Prometheus
and Icarus, Frankenstein, to more modern examples including Neuromancer, Blade
Runner, and Terminator. Id.
180
Id. (“We are fascinated by the ‘black box’ nature of computers and the technical
virtuosity of hackers who manipulate them, but at the same time we are fearful of
their lack of transparency and the fact that our conventional concept of technological
experts may be fatally undermined by largely anonymous, unaccountable, and
potentially subversive technological whiz-kids.”).
181
See generally PHISHING AND COUNTERMEASURES: UNDERSTANDING THE INCREASING
PROBLEM OF ELECTRONIC IDENTITY THEFT (Markus Jakobsson & Steven Myers eds.,
2006) (tracing history of so-called Phishing attacks, in which attackers use email and
fraudulent websites to steal identities).
182
In particular, the fear of anonymity is pronounced. TAYLOR, supra note 29, at
1366
University of California, Davis
[Vol. 41:1327
technology lacks any rational basis183 (after all, technology often fails
in spectacular, frightening ways), the amount of fear and anxiety we
feel about the Internet is difficult to justify.184
The fear of the Internet can trigger the availability heuristic and
probability neglect, explaining in part, but not in whole, the tendency
to exaggerate the power of the Superuser. It might seem odd that
something as intangible and technologically complex as the Internet
would trigger the availability heuristic, which focuses on the familiar
and salient; yet the availability heuristic still exerts great influence for
at least three reasons.
First, although the causes or mechanisms of online risk are
sometimes bogged down in technical detail, they result in tangible,
easy-to-imagine effects. Planes fall out of the sky, power grids go
dark, children are stalked, and credit card numbers are stolen.
Second, the media report on online risks incessantly.185 Third, nearly
everyone has suffered firsthand from some form of online harm.
Computer viruses, hard drive crashes, and spam have become a
constant part of our daily, online existence. Perhaps experiences with
these kinds of minor harms cause people to miscalculate the
likelihood of severe types of online harm. “I am sure hackers can
break into defense department networks,” the reasoning might go,
“because my computer seems to get a new virus every day.”
xiii (“One of the main factors making hacking particularly suitable for media
hyperbole is its aura of anonymity.”).
183
A rich literature describes the technophobia associated with any perceived risk
from new technology. See Levi, supra note 22, at 50.
184
Focusing on public concerns about privacy, Kim Taipale, executive director of
the Center for Advanced Studies in Science and Technology Policy, advances a similar
theory, and points the finger at many sources.
The availability of information privacy horror stories (in particular, the
prevalence of identity theft, spam and hacker stories in the media), and the
general mistrust in government agencies to handle personal information
appropriately, combined with a general apprehension about technology and
how it works, and the natural anxiety relating to disclosure of personal,
particularly intimate, information — all spurred on by the privacy lobby —
has created a public anxiety about electronic privacy out of proportion to the
actual privacy risks and has obscured discussion of the very real threats
posed by either failing to provide security or by misallocating security
resources.
K.A. Taipale, Technology, Security and Privacy: The Fear of Frankenstein, the Mythology
of Privacy, and the Lessons of King Ludd, 7 YALE J.L. & TECH. 123, 137-38 (2004-2005)
(footnotes omitted).
185
See infra Part III.A.2.
2008]
The Myth of the Superuser
1367
Similarly, fear of the Internet triggers probability neglect. Many
harms online are the type of worst case, emotionally charged harms that
Child pornography and
trigger inattention to probabilities.186
cyberterrorism, in particular, supply the public with easy-to-visualize,
horrific, and even apocalyptic imagery.187 As a result of probability
neglect and the availability heuristic, observers imagine there are many
Superusers when in reality there are few, or they believe that Superusers
have a much stronger impact or reach than they actually do.188
2.
Superusers in the Media
It is also useful to examine the relationship between the media and
fear. The media pays a lot of attention to Superusers. This may be a
simple effect of the Myth, or it may be a root cause.
Some sociologists believe the media causes fear.189 According to
these theories, the media exaggerates and overemphasizes stories
about risks to maximize ratings and readership, which exacerbates the
public’s fear.190 Although this so-called “media-effects theory” is
contested,191 regardless of its validity, many agree that the media pays
186
See Cass R. Sunstein, Probability Neglect: Emotions, Worst Cases, and Law, 112
YALE L.J. 61, 76 (2002) (“My central claim is that when strong emotions are involved,
large-scale variations in probabilities will matter surprisingly little — even when the
variations unquestionably matter when emotions are not triggered.”).
187
See Green, supra note 50 (recounting predictions of impending “digital
Armageddon” resulting from cyberterrorism).
188
Another way to characterize this phenomenon involves the types of logical
mistakes caused from the biases of fear. Logicians call this mistake the hasty
generalization or the converse accident. See NICHOLAS BUNNIN & JIYUAN YU, THE
BLACKWELL DICTIONARY OF WESTERN PHILOSOPHY 250 (2004) (explaining entry for
“fallacy of secundum quid”). This informal logical fallacy undermines inductive
reasoning from a particular case to a general rule. Id. When the specific cases are not
numerous enough or typical enough to illuminate the general rule, drawing the latter
from the former is an error. See id.
189
See generally SUNSTEIN, supra note 163, at 87 (“[N]ews sources do a great deal
to trigger fear . . . .”); Ronald Weitzer & Charis Kubrin, Breaking News: How Local TV
News and Real-World Conditions Affect Fear of Crime, 21 JUST. Q. 497, 497 (2004)
(examining “the role of the media in shaping crime fears, in conjunction with both
demographic factors and local crime conditions”).
190
LORI DORFMAN & VINCENT SCHIRALDI, BUILDING BLOCKS FOR YOUTH, Off Balance:
Youth,
Race
&
Crime
in
the
News
(2001),
available
at
http://www.buildingblocksforyouth.org/media/media.html (describing how media
portrayals of crime can drive public policy responses and frame issues for public).
191
See BARRY GLASSNER, THE CULTURE OF FEAR: WHY AMERICANS ARE AFRAID OF THE
WRONG THINGS 26-29 (1999); Karen Frost et al., Relative Risk in the News Media: A
Quantification of Misrepresentation, 87 AM. J. PUB. HEALTH 842, 844 (1997); David
Gauntlett, Ten Things Wrong with the “Effects Model” (1998), available at
1368
University of California, Davis
[Vol. 41:1327
disproportionate attention to sensational stories aimed at fearinducing topics.192 The Superuser is an example: he is a pervasive
image, especially compared to the exaggerated focus on online crimes
committed by unsophisticated non-Superusers.
News stories about online threats appear regularly in major
newspapers.193 Many of these stories exaggerate the sophistication of
the crimes and the criminals. Consider, for example, convicted hacker
Kevin Mitnick. By some accounts, Mitnick is the most notorious
computer hacker in the world.194
Mitnick’s forte was “social
engineering,” a glorified term for skillful lying.195 He once obtained
proprietary source code and manuals from a Pacific Bell field office by
convincing the person at the front desk of a data center that he was an
employee.196
Most of his infamous attacks relied on social
engineering, not technical wizardry.197
http://www.theory.org.uk/david/effects.htm.
192
E.g., DORFMAN & SCHIRALDI, supra note 190 (surveying various print and
broadcast news sources spanning nearly 100 years and concluding “[t]he news media
report crime, especially violent crime, out of proportion to its actual occurrence”).
193
Consider the following headlines that appeared in the New York Times in 2006:
Associated Press, Computer Hackers Attack State Department, N.Y. TIMES, July 12, 2006,
at A6; Associated Press, Hackers Gain Data on AT&T Shoppers, N.Y. TIMES, Aug. 30,
2006, at C2; William L. Hamilton, You’re Not Alone, N.Y. TIMES, Nov. 23, 2006, at F1
(describing threat to networked home computers as “the next frontier of risk”); Metro
Briefing/New Jersey, Newark: University Computers Hacked, N.Y. TIMES, Apr. 10, 2006,
at B4; Alex Mindlin, Your Computer Is Under Attack — LOL, N.Y. TIMES, Feb. 20, 2006,
at C3; David Shenk, A Growing Web of Watchers Builds a Surveillance Society, N.Y.
TIMES, Jan. 25, 2006, at G6; Tom Zeller, Jr., Cyberthieves Silently Copy Your Passwords
as You Type, N.Y. TIMES, Feb. 27, 2006, at A1.
194
See Michael Specter, An Ex-Con Logs On, NEW YORKER, Feb. 3, 2003, at 32
(stating “Mitnick . . . is usually described as the world’s most notorious hacker”);
Patricia Jacobus, Mitnick Released from Prison, CNET NEWS.COM, Mar. 24, 2001,
http://www.news.com/Mitnick-released-from-prison/2100-1023_3-235933.html
(noting “Kevin Mitnick, one of the world’s most notorious computer hackers”).
195
In his post-prison, reformed public persona, Mitnick has even written a few
books about social engineering. See, e.g., KEVIN D. MITNICK & WILLIAM L. SIMON, THE
ART OF DECEPTION: CONTROLLING THE HUMAN ELEMENT OF SECURITY (2002) (exploring
how hackers use social engineering); KEVIN D. MITNICK & WILLIAM L. SIMON, THE ART
OF INTRUSION: THE REAL STORIES BEHIND THE EXPLOITS OF HACKERS, INTRUDERS &
DECEIVERS (2005) (describing real-life stories of computer intrusions, many involving
social engineering).
196
See KATIE HAFNER & JOHN MARKOFF, CYBERPUNK 50-51 (1991).
197
See id.; Elizabeth Weise, Hacker Prowess Exaggerated? Computer Villain Seen as
Virtually Marginal in Reality, CHARLOTTE OBSERVER, Jan. 27, 1996, at 2A (explaining
that “in news reports, Mitnick was pictured as a lone, master hacker, capable of doing
almost anything with a computer or even just a phone. In reality, Mitnick’s technical
skills were only fair”).
2008]
The Myth of the Superuser
1369
Despite the low-tech methods Mitnick used, some in the media have
portrayed him as a technical genius.198 A New York Times article
written at his arrest breathlessly announced that “[t]he technical
sophistication of the pursued and his pursuer . . . was remarkable.”199
In addition to these newspaper accounts, authors have written many
books about the lore of the Superuser hacker.200 The pervasive
attention to the Superuser extends beyond print media. The lone,
genius hacker has become almost a stock figure in many movies, such
as WarGames, The Matrix, TRON, and Sneakers.201
The media’s exaggeration of technical sophistication is a shame,
because careful media attention can dispel the Myth and provide a
calming influence on public fears. For example, three praiseworthy
articles in the New York Times in 2006 sought to defuse, not heighten,
the Myth of the Superuser. One story described the realization by
officials at ChoicePoint, a prominent data broker, that despite all of its
efforts to harden its computer databases, unscrupulous “customers” were
buying information under the cover of legitimate businesses.202 Another
198
See Associated Press, Cyberspace Raider to Get Plea Bargain, ATLANTA J. &
CONST., July 2, 1995, at A5 (calling Mitnick “[a] computer hacker with a history of
breaking into some of the nation’s most protected computer systems”); Bernard Levin,
Misappliance of Science, TIMES (U.K.), Mar. 24, 1995, at 16 (listing crimes attributed to
Mitnick to conclude that “this man is a genius,” and “Mitnick, it is very clear, could
clean out a thousand bank systems and retire with countless billions of dollars”).
199
John Markoff, Hacker and Grifter Duel on the Net, N.Y. TIMES, Feb. 19, 1995, at
A1. By “pursuer,” Markoff is referring to Tsutomo Shimomura, a computer researcher
who helped find Mitnick. Granted, the article goes on to mention that “[i]f anything,
Mr. Mitnick’s real ‘darkside’ brilliance comes not from his computer skills, but from
his insight into people.” Id. After Mitnick’s arrest, Markoff and Shimomura coauthored a book, JOHN MARKOFF & TSUTOMO SHIMOMURA, TAKE-DOWN: THE PURSUIT
AND CAPTURE OF KEVIN MITNICK, AMERICA’S MOST WANTED COMPUTER OUTLAW — BY
THE MAN WHO DID IT (1996). Reportedly the marketing copy for the book referred to
Mitnick as “a hacker who ‘could have crippled the world.’” Greg Miller, Did Reporter
Sensationalize Case?, SEATTLE TIMES, Mar. 20, 1999, at A2.
200
See generally HAFNER & MARKOFF, supra note 196 (profiling several famous
computer hackers); JONATHAN LITTMAN, THE FUGITIVE GAME: ONLINE WITH KEVIN
MITNICK (1996) (describing pursuit and arrest of Mitnick); MARKOFF & SHIMOMURA,
supra note 199 (same); BRUCE STERLING, THE HACKER CRACKDOWN (1992) (describing
arrest of Mitnick); CLIFFORD STOLL, THE CUCKOO’S EGG (1989) (detailing pursuit of
computer intruders).
201
Others deemed merely footnote-worthy include The Matrix sequels (Warner
Bros. 2003), HACKERS (Metro-Goldwyn-Mayer 1995), THE NET (Columbia Pictures
1995), SWORDFISH (Warner Bros. 2001), JOHNNY MNEMONIC (Sony Pictures Home
Entertainment 1995), GOLDENEYE (Metro-Goldwyn-Mayer 1995), THE LAWNMOWER
MAN (New Line Cinema 1992), EXISTENZ (Alliance Atlantis 1999), and ANTITRUST
(Metro-Goldwyn-Mayer 2001).
202
Gary Rivlin, Keeping Your Enemies Close, N.Y. TIMES, Nov. 12, 2006, § 3 at 1.
1370
University of California, Davis
[Vol. 41:1327
article tried to spotlight the difference between “data breach” and
“identity theft,” concepts that are often conflated by the media.203 A
third article argued that when it comes to stolen data and identity theft,
“[h]ackers and sophisticated data thieves are one thing. But in the battle
to stop the great hemorrhaging of personal data, the enemy is us.”204
Thus, conventional social science explains many of our observations
about the Myth of the Superuser. Fear, abetted by the media spotlight,
causes people to rely on the availability heuristic and to suffer from
probability neglect. Both phenomena lead to exaggerated risk
assessments.
B. Myth Origins: Technology
Heuristics and biases explain why we exaggerate the online fears we
have, but they cannot so easily explain why exotic, new, online fears
develop so rapidly, and why the number and variety of myths seem so
plentiful compared to other fields. Why are new fears added so
quickly to the Internet’s fear-processing machine? The way to account
for this is to look at the nature of the technology; networks, software,
and hardware are malleable, but only to an extent. Due to confusion
about Lessig’s ideas about code, this malleability is too often mistaken
for boundless possibility.
1.
Malleability, or the Care and Feeding of Superusers
Superusers thrive by taking advantage of several well-known
features of programmers, code, computers, and networks. First,
Superusers benefit from the openness of software and hardware.205
Computer hardware is almost always shipped in an easy-to-open case
that invites tinkering, even though most computer users will never
tinker.206 Computer software is sold in a metaphorically similar
manner, with the typical operating system (“OS”) shipped to allow
“administrator access” by the average user.207
203
Steve Lohr, Surging Losses, But Few Victims, N.Y. TIMES, Sept. 27, 2006, at G1.
Tom Zeller, Jr., 93,754,333 Examples of Data Nonchalance, N.Y. TIMES, Sept. 25,
2006, at C5.
205
See Zittrain, supra note 156, at 1982-87.
206
See generally WINN L. ROSCH, THE WINN L. ROSCH HARDWARE BIBLE (6th ed.
2003) (providing detailed guide for repairing and understanding personal computer
hardware).
207
See Zittrain, supra note 156, at 1983 (“Most significant, PCs were and are
accessible. They were designed to run software not written by the PC manufacturer or
OS publisher, including software written by those with whom these manufacturers
204
2008]
The Myth of the Superuser
1371
These are not mandatory design choices. Hardware could be
shipped sealed and inaccessible, and the OS could allow only limited
control.208 Were those choices the status quo, it would be harder to be
a Superuser.209 This is why experts modify and adapt the open PC
much more easily than the closed, hard-to-modify TiVo.210
Second, networks are also open to scrutiny and manipulation.211
Although openness is why the Internet has so grown rapidly to
include many innovative services, it can be exploited by the Superuser.
For example, robust authentication controls — mechanisms to verify
that a person online is who they say they are — were not designed
into the Internet’s core protocols.212 Although authentication has been
added after the fact, the unauthenticated core is always there.213 Thus,
Superusers can take advantage of the network’s trust to act undetected.
Third, software will always be imperfect. All commercial software
programs have bugs214 because it would be too expensive to drive
them all away.215
Superusers find and exploit these bugs to
had no special arrangements.”).
208
Microsoft has trumpeted the fact that its latest version of Windows, Vista,
makes nonadministrator access the default. See Posting of Jim Allchin to Windows
Vista Team Blog, http://windowsvistablog.com/blogs/windowsvista/archive/2007/
01/23/security-features-vs-convenience.aspx (Jan. 23, 2007, 17:32).
209
See Zittrain, supra note 156, at 1982-87; cf. Jay Kesan & Rajiv C. Shah, Setting
Software Defaults: Perspectives from Law, Computer Science and Behavioral Economics,
82 NOTRE DAME L. REV. 583 (2006) (explaining how defaults operate in software and
how policymakers should set defaults).
210
See Zittrain, supra note 156, at 2014-15.
211
See Mark Lemley & Lawrence Lessig, The End of End-to-End: Preserving the
Architecture of the Internet in the Broadband Era, 48 UCLA L. REV. 925, 930 (2001) (“It
is the view of many in the Internet community, ourselves included, that the
extraordinary growth of the Internet rests fundamentally upon its design principles.
Some of these principles relate to the openness of the Internet’s standards and the
openness of the software that implemented those standards.”).
212
Cf. KIM CAMERON, THE LAWS OF IDENTITY (2005), available at
http://www.identityblog.com/stories/2005/05/13/TheLawsOfIdentity.pdf (discussing
why it is hard to add identity to Internet).
213
See David Talbot, The Internet is Broken, TECH. REV., Dec. 20, 2005, available at
http://www.technologyreview.com/Infotech/16051/page2/ (“Simply put, the Internet
has no inherent security architecture — nothing to stop viruses or spam or anything
else. Protections like firewalls and antispam software are add-ons, security patches in
a digital arms race.”).
214
Crispin Cowan, Calton Pu, & Heather Hinton, Death, Taxes, and Imperfect
Software: Surviving the Inevitable, in ACM PROCEEDINGS OF THE 1998 WORKSHOP ON
NEW SECURITY PARADIGMS 54, 55 (1998) (“Commercial software chronically has bugs,
many with security vulnerability implications.”).
215
See id. “Tempting as it may be to hypothesize that this is because the vendors
are lazy or stupid, this is not the case. Commercial software chronically has bugs for
1372
University of California, Davis
[Vol. 41:1327
circumvent security, break DRM, or otherwise cause software to do
what it is not designed to do.216
Finally, programmers — often Superusers themselves —
purposefully enable expert level control frequently. Consider, for
example “command line” programs such as UNIX shells or the
Windows command prompt.217 With these programs (which are quite
homely by today’s graphical standards), users key in esoteric
commands to control the OS, for example, to copy files, create folders,
or run programs.218 Although they can do these things with a modern
graphical user interface, experienced command line users consider
themselves more efficient than their mouse-bound counterparts.219
2.
Misunderstanding “Code Is Law”
Relying on Lessig’s important work on code for support, people
mistakenly interpret the Internet’s malleability to mean limitless
possibility for Superuser power.220 Lessig famously spotlighted the role
that software plays in regulating online behavior.221 Easily modifiable
software, Lessig noted, defines the “laws of nature” of online spaces.222
If a programmer decides he wants people to “walk through walls”
online, he can change the code to allow that behavior. In this way,
software regulates online conduct in much the same way laws (and
norms and markets) do.223 In slogan form, “Code is Law.”224
the dual-reason that correctness is hard, and correctness does not sell software.” Id.;
see also Eric Sink, Why We All Sell Code with Bugs, GUARDIAN UNLIMITED, May 25,
2006, at 6, available at http://www.guardian.co.uk/technology/2006/may/25/
insideit.guardianweeklytechnologysection.
216
Brad Stone, A Lively Market, Legal and Not, for Software Bugs, N.Y. TIMES, Jan.
30, 2007, at A1 (describing “the willingness of Internet criminals to spend large sums
for early knowledge of software flaws that could provide an opening for identity-theft
schemes and spam attacks”).
217
See NEAL STEPHENSON, IN THE BEGINNING . . . WAS THE COMMAND LINE 13 (1999)
(describing command lines).
218
See JERRY PEEK, GRACE TODINO, & JOHN STRANG, LEARNING THE UNIX OPERATING
SYSTEM 11-14 (5th ed. 2002) (describing basic UNIX commands).
219
See STEPHENSON, supra note 217, at 74. Stephenson describes how Apple
programmers created a command line interface on the early Macintosh computer —
the symbol of the birth of the graphical user interface — “so that they would be able
to get some useful work done.”
220
See infra notes 225-27 and accompanying text.
221
LESSIG, supra note 7 passim.
222
Id. at 24.
223
See id. at 125.
224
Id. at 1.
2008]
The Myth of the Superuser
1373
Unfortunately, few have accepted Lessig’s important and useful
insights as an invitation to study the malleability of code in a rigorous
way. Legal academics, in particular, have embraced the general idea
that code is important, but they too often treat it as mysterious and
complex — a hopeless moving target.225 The tendency is to hope that
code will evolve to resolve conflicts, an unrealistic technological
determinism.226 Worse, others fall into a “science fiction trap,”
imagining that every kind of new technology is possible.227 By
ignoring the constraints of reality, these people can make any problem
melt away. The truth, however, is not so rosy.
In particular, the misinterpretation of Lessig’s ideas has led to what I
call “metaphor failure.” Internet law is often a battle of metaphors.
When reading my email messages, is my ISP acting more like the
postman who glances at the backs of postcards or the one who rips
225
See R. Polk Wagner, On Software Regulation, 78 S. CAL. L. REV. 457, 492 (2005)
(“[S]oftware development is a rapidly moving, nearly unpredictable target, making it
difficult to directly address software through legal regulation.”). I do not mean to be
too critical of Professor Wagner’s article, because although I think he overemphasizes
the unpredictability and instability of code development, the article stands as a leading
example of a careful analysis of law’s effect on code. His overemphasis on the
instability of code causes him to offer prescriptions that are much too deferential, but
this is a critique of fine points while celebrating the article’s larger goals.
226
There is an ancient (in network terms) famous mantra that exemplifies this
point: “[T]he [Inter]net interprets censorship as damage and routes around it.” See
GOLDSMITH & WU, supra note 6, at 3 (quoting John Gilmore). Except sometimes the
Internet does not route around censorship, and censorship flourishes. James Fallows,
‘The Connection Has Been Reset,’ ATLANTIC MONTHLY, Mar. 2008, available at
http://www.theatlantic.com/doc/200803/chinese-firewall (describing China’s use of
firewalls and other Internet technology to monitor and block disfavored traffic).
227
Lessig himself has occasionally been guilty of falling for the science fiction trap,
most conspicuously by placing too much stock in the Platform for Privacy Preferences
(“P3P”). P3P is a form of so-called “intelligent agents,” little pieces of software which
might automate and mediate our privacy wishes with the websites we visit. In Code,
Lessig argued that “with a technology like P3P, we could lower transaction costs
enough to make a property rule [instead of a liability rule] work.” LESSIG, supra note 7,
at 229. As Lessig himself noted in the 2006 version of the book, his recommendation
to propertize privacy has been criticized. Id. at 383 n.47 (citing critics). Lessig’s vision
of the P3P protocol — easy to use, easy to deploy, rich enough to contain our
preferences but simple enough to be understandable — assumes away many difficult
technical problems. See ELEC. PRIVACY INFOR. CTR & JUNKBUSTERS, PRETTY POOR
PRIVACY:
AN ASSESSMENT OF P3P AND INTERNET PRIVACY (2000),
http://epic.org/reports/prettypoorprivacy.html (“[P3P] is a complex and confusing
protocol that will make it more difficult for Internet users to protect their privacy.”).
1374
University of California, Davis
[Vol. 41:1327
open closed envelopes?228 Is an encrypted document more like a
paper letter inside a closed box or a shredded document?229
Superusers’ actions seem more science fiction than reality. As one
scholar noted, “What is talked about, in terms of hackers at least, is
the manner in which hackers themselves exist in a shadowy space of
secrecy, possessing near mystical powers that allow control of
technology that itself is beyond discourse.”230 A hacker can pass
through “impenetrable” firewalls231 (walk through walls), install a
rootkit232 (leave behind no trace), scan entire networks in search of
interesting files in a matter of minutes233 (fly through entire
neighborhoods of information), and walk off with millions of
identities (thousands of pages of information) never to be heard from
again (vanish).
When metaphors for online concepts fail, scholars and policymakers
become deeply unmoored. Stripped of comparison points, they see
online conflicts as blank slates. These conflicts provide opportunities
to rewrite rules and start from scratch, propose creative and untested
solutions, and abandon ordinary tools that have been used for decades
in real world conflicts. This is a form of an Internet exceptionalist
strain of thinking that many scholars have debunked in recent years
but that stubbornly persists among policymakers and even some
228
See Orin S. Kerr, The Problem of Perspective in Internet Law, 91 GEO. L.J. 357,
365-68 (2003).
229
See A. Michael Froomkin, The Metaphor Is the Key: Cryptography, the Clipper
Chip, and the Constitution, 143 U. PA. L. REV. 709, 884 (1995).
230
Thomas, supra note 114, at 27.
231
See Byron Acohido & Jon Swartz, Unprotected PCs Can Be Hijacked in Minutes,
USA TODAY, Nov. 29, 2004, at 3B (finding that “firewalls, which restrict online access
to the guts of the PC operating system, represent a crucial first line of defense against
cyberintruders”).
232
Rootkits are programs, typically installed by computer intruders that wipe out
traces of the intruder’s identity from the computer’s security detection systems. See
Paul Roberts, RSA: Microsoft on ‘Rootkits’: Be Afraid, Be Very Afraid, COMPUTERWORLD,
Feb. 17, 2005, available at http://www.computerworld.com/securitytopics/security/
story/0,10801,99843,00.html (describing new, more threatening rootkit technologies).
233
See J.D. Biersdorfer, Q&A: From a Crisp Web Image to a Blur, via the Printer,
N.Y. TIMES, Sept. 6, 2001, at G4.
A port scan means that another computer on the Internet is looking for an
open door to your machine. Port scans are legal and are used in some cases
for network management and administration, but hackers are also
increasingly using port scanning to find a way to break in so that they can
tamper with the computer or steal data from it.
Id.
2008]
The Myth of the Superuser
1375
academics.234 To these people, imagining the possibility of a harm is
enough; confirming such speculation with people with firsthand (or
even secondhand) experience is optional.
We are better off once we realize that although code is malleable, it
is also constrained and predictable. Code can and will change, but at
least within a short time frame, it will only do so in certain ways.
Mindful of the many constraints on software, we can make better
predictions, and avoid and dispel misguided thoughts about the
endless power of the Superuser.
3.
The Truth About Code: Constraints, Not Endless Possibility
There are many constraints on the evolution and development of
software. First, there are constraints of technology. Software can do
only what is permitted by hardware and other software. Programs
cannot interact with the physical world in futuristic, fantastic ways
unless a piece of hardware facilitates that type of interaction.235
Application programs cannot manipulate files unless the OS allows it
to do so.236
Of course, given the generativity of computers, constraints of
technology are surmountable with time and money.237 This suggests
the second, closely related constraint: the constraint of organization.
Despite the romantic vision of the lone programmer, toiling away on
the “Great American Program,” most of the software we use every day
was written by a large committee.238 Although wealthy, large
corporations can finance complex products, their resources are not
234
See, e.g., Timothy Wu, When Code Isn’t Law, 89 VA. L. REV. 679, 682 (2003);
James Grimmelman, Note, Regulating by Software, 114 YALE L.J. 1719, 1728-30 (2005).
I also made this point in a student note. Paul Ohm, Note, Usenet: On Regulating the
Internet, 46 UCLA L. REV. 1941, 1957 (1999).
235
It is not enough to simply have the right piece of hardware. Some hardware
will be constrained, again by software. For example, device drivers are small pieces of
critically important software that allow an operating system to interact with specific
bits of hardware. See Zittrain, supra note 156, at 2018.
236
See NEMETH ET AL., supra note 9, at 73-74 (describing UNIX operating system’s
file permissions methods).
237
See Zittrain, supra note 156 passim (using label “generativity” to describe
malleability of hardware and software).
238
See FREDERICK P. BROOKS, THE MYTHICAL MAN MONTH: ESSAYS ON SOFTWARE
ENGINEERING 31 (20th anniv. ed. 1995) (reporting that at peak, 1000 people worked
simultaneously to support creation of Operating System OS/360); cf. STEVEN WEBER,
THE SUCCESS OF OPEN SOURCE 59 (2004) (noting that most software written by one
person “is used only by the author or perhaps a few friends”).
1376
University of California, Davis
[Vol. 41:1327
limitless; internal politics, inertia, and other similar “frictions” also
limit the type and pace of innovation they are likely to achieve.239
The market serves as another organizational constraint. Although
small tools are often developed to scratch a particular developer’s
idiosyncratic itch, larger products often need a market to emerge.240
Market forces channel developers away from the outlandishly new in
favor of gradual change instead.
Finally, software is written by people who toil under constraints of
human fallibility.
People tend to be busy, distracted, and
unorganized.241 They make mistakes and often lack imagination.242
Witness the web browser. Despite more than a decade of innovation,
the web browser continues to impart information essentially as it did
at its invention with pages full of text and graphics (and in the
principal subsequent innovation, the occasional video clip) scrolling
up and down the screen.243
The limits of human bandwidth serve as another form of human
fallibility, one that is perhaps even more important than the failure of
imagination. The average programmer can keep only a small amount
of code straight in her head and thus can produce only a modest
amount of new code daily.244 Consequently, a finite number of
programmers — for example, Superuser DRM circumventors — can
239
See WEBER, supra note 238, at 57-65 (describing “problem” of developing
complex software). Even Open Source software is subject to this constraint. Any
reasonably complex, widely used Open Source project — Linux, Firefox, Apache,
MySQL — required the work of dozens or more programmers to achieve the stability
and feature set of a mature product. See id. at 94-127 (describing development of
Linux, Apache, and Mozilla). Tales from the trenches describe internecine battles and
outright schisms (explaining code fork, to use a technical phrase) in the development
of some of these products. Id. at 64.
240
But see YOCHAI BENKLER, THE WEALTH OF NETWORKS: HOW SOCIAL PRODUCTION
TRANSFORMS MARKETS AND FREEDOM 41-48 (2006) (describing increasing importance of
nonmarket drivers in information industries).
241
Cf. BROOKS, supra note 238, at 30 (reporting statistics about difference in
productivity between best and worst computer programmers).
242
William S. Curran, The Outer Limits of Programming, in PROCEEDINGS OF THE
38TH ACM SOUTHEAST REGIONAL CONFERENCE, 2000, at 38, 38 (commenting that
human computer programmers “have limits to our mental capacities”).
243
See Andrew Orlowski, Browser Innovation is Dead — Andreessen, THE REGISTER,
July 2, 2003, http://www.theregister.co.uk/2003/07/02/browser_innovation_is_dead_
andreessen/ (quoting Marc Andreessen, inventor of Mosaic and Netscape browsers,
“there hasn’t been any innovation on the browser in the last five years”).
244
Curran, supra note 242, at 38 (noting that 50 lines of code is both supposed
“practical limit [of] the number of lines [of programming code] that a typical
programmer can keep clearly in his/her head” as well as upper bound of what “average
programmer produces . . . daily”).
2008]
The Myth of the Superuser
1377
produce only a finite output.245 Whether a particular new DRM
scheme is attacked and successfully circumvented, then, depends on
the aggregate priorities of a number of people.
The power of these constraints varies depending on the interval of
time about which predictions are being made. Over a long time
horizon — say ten or twenty years — these constraints will have
minimal effect on the evolution of software. Over the short term,
these constraints dominate, and software tends to evolve, not lurch.246
If we view claims of Superuser power through the lens of these
constraints — technology, organization, and human fallibility — we
can separate the probable from the merely possible. In a world of
endless possibility, Superusers can do anything online. But in reality,
many would-be Superusers will try to hack software and fail, lack the
imagination or expertise to succeed, or look for co-conspirators and
find none. Then having failed, they will move on to the next project,
other diversion, or day job.
C. Myth Perpetuation: The Failure of Expertise
What I have described so far is a fear-processing machine fed with
stories about all-powerful Superusers, supplied at an amplified rate
due to misunderstandings about the malleability of code, and acted
upon by well-known heuristics and biases we suffer in the face of fear.
The final piece is the machine’s operator. Most of Sunstein’s examples
involve risks with well-known probabilities calculated by experts.247
These experts regulate the knobs and dials of the machine, trying to
slow the output to a manageable rate. The problem with online risk is
that even experts have trouble assigning meaningful probabilities to
different risks because the machine is unattended. How often will a
hacker successfully breach a server? How often does unlawful access
245
Id. (“Now let’s consider a program estimated to require, say, ten million lines of
code. At ten lines per day, that would take a thousand programmers a thousand days.
That’s about four years of work for the thousand programmers.”).
246
In the terms of evolutionary biology, software evolves through phyletic
gradualism, not punctuated equilibrium. Cf. THOMAS KUHN, THE STRUCTURE OF
SCIENTIFIC REVOLUTIONS (1962) (introducing concepts of paradigm shifts); Niles
Eldredge & Stephen Jay Gould, Punctuated Equilibria: An Alternative to Phyletic
Gradualism, in MODELS IN PALEOBIOLOGY 82 (Thomas J.M. Schopf ed., 1972)
(introducing
theory
of
punctuated
equilibrium),
available
at
http://www.blackwellpublishing.com/ridley/classictexts/eldredge.asp.
247
E.g., SUNSTEIN, supra note 163, at 96-97 (discussing risk of contracting SARS in
Canada); id. at 117 (discussing risks of harm from cell phone radiation and arsenic in
drinking water); id. at 132-36 (describing various calculations of value of statistical life).
1378
University of California, Davis
[Vol. 41:1327
to a social security number result in identity theft? Experts rarely ask
these types of questions.
Contrast Sunstein’s paradigmatic experts: scientists who assess the
risks of cancer, the safety of nuclear power plants, or the dangers of
genetically modified foods.248 These researchers rigorously analyze
statistics to calculate risks.249 In fact, the disconnect between their
focus on numbers and probabilities and the average person’s seeming
disregard for statistics is a central mystery pursued in Laws of Fear.250
In stark contrast, experts in the field of computer crime and
computer security seem uninterested in probabilities. Why are experts
in this field willing to abdicate the important risk-calculating role
played by their counterparts in other fields? Consider the following
four explanations.
1.
Pervasive Secrecy
Online risks are shrouded in secrecy. Software developers use trade
secret law and compiled code to keep details away from public
consumption.251 Computer hackers dwell in a mythical, shadowy
underground and trade vulnerabilities in private chat rooms.252
Security consultants are contractually bound not to reveal the
248
Id. at 139 (describing studies on eliminating cancer risks); id. at 47 (discussing
expert evaluations of risk from nuclear power plants); id. at 40 (discussing concerns
about genetically modified organisms).
249
Id. at 139.
250
This is not to say that Sunstein and the social scientists he cites find all expert
risk assessment flawless. At one point, Sunstein concedes that experts tend to “use
their own heuristics and have their own biases.” See id. at 86-87 (citing SHELDON
RAMPTON & JOHN STAUBER, TRUST US, WE’RE EXPERTS! (2001)). He does not, however,
spend much time on this observation, nor does he point out any situations in which
experts seem to have the very same biases as nonexperts. See also Kahneman &
Tversky, supra note 170, at 18 (explaining that “[t]he reliance on heuristics and the
prevalence of biases are not restricted to laymen . . . [e]xperienced researchers are also
prone to the same biases — when they think intuitively”).
Kahneman and Tversky (and probably Sunstein) seem to be talking about
exceptional instances where experts fall prone to biases despite their training and
rigor. The studies they cite, for example, focus on problems that are “more intricate”
and “less transparent” than the “elementary errors” made by laypeople. Id. at 18. In
contrast, I claim that experts of online risk tend to make precisely the same mistakes
of judgment as made by laypeople.
251
See WEBER, supra note 238, at 192. “[C]ontrol of the source code is the
foundation of [the traditional software] business model. . . . The simplest way to
retain control is to give only the binary executable codes to the customer.” Id. The
obvious exception to this tendency is the Open Source model. Id.
252
See Thomas, supra note 114, at 27.
2008]
The Myth of the Superuser
1379
identities of those who hire them.253 Law enforcement agencies refuse
to divulge statistics about the number, type, and extent of their
investigations and resist congressional attempts to increase public
reporting.254
Since the September 11 attacks, the secrecy culture surrounding
code has broadened. The government has introduced new secrecy
measures with the goal of protecting our nation’s communications
networks, now part of the governmentally defined “Critical
Infrastructure.”255 For example, industry experts and government
officials meet periodically to share information about online risks in a
so-called Information Sharing and Analysis Center; those discussions
are not made available to the public.256
2.
Everybody Is an Expert
The world abounds with computer experts, simply because theirs is
a title too easily obtained. All you need is skill with a computer;
formal training and background in rigorous methods are not required.
In fact, to many of these anointed experts, most academic types —
253
See Levi, supra note 22, at 46.
When Congress passed the update to the USA PATRIOT Act, known as the USA
PATRIOT Improvement and Reauthorization Act of 2005, it sought to increase
congressional oversight, by requiring additional auditing and reporting to Congress
about the government’s uses of certain FISA authorities. USA PATRIOT Improvement
and Reauthorization Act of 2005, Pub. L. No. 109-177, §§ 106, 119, 120 Stat. 278,
199-200 (2006). President George W. Bush issued a signing statement to the new law,
reserving the right not to comply with the new requirements. President George W.
Bush, President’s Statement on H.R. 199, the “USA PATRIOT Improvement and
Reauthorization Act of 2005” (Mar. 9, 2005), available at http://www.whitehouse.gov/
index.html (search “President’s Statement on H.R. 199, the ‘USA PATRIOT
Improvement and Reauthorization Act of 2005’”; then click on first hyperlink). The
statement said, in pertinent part:
254
The executive branch shall construe the provisions of H.R. 3199 that call for
furnishing information to entities outside the [E]xecutive [B]ranch, such as
sections 106A and 119, in a manner consistent with the President’s
constitutional authority to supervise the unitary executive branch and to
withhold information the disclosure of which could impair foreign relations,
national security, the deliberative processes of the Executive, or the
performance of the Executive’s constitutional duties.
Id.
255
See NATIONAL STRATEGY, supra note 86, at 13-14.
See Kevin Poulsen, Feds Urge Secrecy over Network Outages, SECURITYFOCUS,
June 23, 2004, http://www.securityfocus.com/news/8966 (“Data exchanged within the
Telecom-[Information Sharing and Analysis Center] is protected from public
disclosure.”).
256
1380
University of California, Davis
[Vol. 41:1327
those likeliest to conduct rigorous risk assessments — are, in fact,
inexpert because they focus too much on theory over practice.257
Part of this stems from the garage hacker history of computer
innovation.258 Unlike modern medicine, where most important
advances require money and years of formal education to achieve,
many computer breakthroughs come from self-taught tinkerers.259
Generally, this democratizing nature of online expertise would be
cause for celebration.
The problem is that self-educated computer experts tend to have
neither the training nor inclination to approach problems statistically
and empirically. People may be called before Congress to testify about
identity theft or network security, even if they have no idea nor even
care how often these risks occur.260 Their presence on a speakers’ list
crowds out the few academics who are thinking about these things
empirically and rigorously.261
257
For example, in reporting news that Microsoft had opened a new research
facility in Cambridge, Massachusetts, the New York Times quoted many sources that
suggested a divide between theoretical computer science and “useful” “product
development.” Katie Hafner, Microsoft Adds Research Lab in East as Others Cut Back,
N.Y. TIMES, Feb. 4, 2008, at C3; cf. Robert L. Glass, Revisiting the Industry/Academe
Communication Chasm, 40 COMM. ACM 11, 13 (1997) (fearing “communication chasm
between [computer science] academe and industry” which are “unnecessary because
both academics and practitioners are typically bright and rational people who share
similar goals”).
258
See PAUL FREIBERGER & MICHAEL SWAINE, FIRE IN THE VALLEY: THE MAKING OF
THE PERSONAL COMPUTER 78-79, 118-24 (2d ed. 1999) (describing role of hobbyists
and enthusiasts in establishing market for PCs); Zittrain, supra note 156, at 1984-85
(discussing role of hobbyists in development of commercial software); Posting of
Howard Rheingold to Huridocs-Tech, Human Rights Education Associates, (Dec. 23,
1999), available at http://www.hrea.org/lists/huridocs-tech/markup/msg00383.html.
259
See Walter Isaacson, Thinkers vs. Tinkerers, and Other Debates, TIME, Mar. 29,
1999, at 6 (comparing “relative influence of thinkers vs. tinkerers” over time, and
focusing in particular on role of tinkerers in computing advances).
260
For example, a subcommittee of the House Judiciary Committee recently held a
hearing to discuss identity theft. Privacy and Cybercrime Enforcement Act of 2007:
Hearing on H.R. 4175 Before the Subcomm. on Crime, Terrorism, and Homeland Security
of the H. Comm. on the Judiciary, 110th Cong. (2007), available at
http://judiciary.house.gov/Hearings.aspx?ID=192. The witness list consisted of three
government officials, a victim of identity theft, the president of a software industry
association, and a privacy advocate. Id. While many of these witnesses reported
statistics about the number of identity theft cases each year, none tried to compare the
risks of identity theft with other online or offline risks. E.g., id. (statement of Craig
Magaw, Special Agent in Charge, Criminal Investigative Division, U.S. Secret Service)
(reporting that Secret Service agents had arrested over 4300 suspected identity thieves
in fiscal year 2007).
261
See infra Part IV.B.2.b for a description of some academic disciplines attempting
2008]
The Myth of the Superuser
1381
This may be part of a broader trend. Professor Suzanna Sherry has
written about a tendency in law and policy to shift away from relying
on expert knowledge to consulting mass-created, democratic
knowledge.262 She points to prominent legal scholars in constitutional
law, administrative law, and civil procedure who have called for this
shift,263 often in reaction to what they see as the excesses of the power
of unelected judges.264 Sherry is a strong critic of this trend, referring
by analogy to two disastrous experiments with the politicization of
scientific inquiry: (1) the Soviet Communist party’s endorsement of
dubious genetic theories known as Lysenkoism and (2) the debate
over evolution and creationism in American public schools.265
If Sherry is right, then perhaps the devaluation of expertise we see
with computers will soon spread to other scientific and engineering
disciplines that raise public fears, such as environmental science,
genetic engineering, medicine, and biotechnology. Already vulnerable
to miscalculated risks, if these fields begin to lose their scientific
gatekeepers, the problems of biases and heuristics will worsen and
become more difficult to remedy.
3.
Self-Interest
Many people are selfishly motivated to portray online actors as
sophisticated hackers capable of awesome power.266 Prosecutors
characterize criminal defendants as evil masterminds to gain jury
appeal or to enhance a sentence.267 Law enforcement officials spin
to bring rigor to online risk assessments.
262
Suzanna Sherry, Democracy and the Death of Knowledge, 75 U. CIN. L. REV. 1053,
1053-54 (2007).
263
See id. at 1057 n.10.
264
See id. at 1057 n.9 (citing ALEXANDER M. BICKEL, THE LEAST DANGEROUS BRANCH:
THE SUPREME COURT AND THE BAR OF POLITICS 16 (1962)).
265
See id. at 1067-69. Her principal focus is not on scientific expertise, but instead
on the “democratization” of legal knowledge, and in particular constitutional
interpretation.
266
Paul Taylor has remarked:
Despite their diametrically opposed arguments as to the potential social and
technical benefits of hacking, both computer security figures and denizens of
the computer underground occasionally manipulate and exaggerate the
malevolent aspects of hacking (the former to stigmati[z]e and isolate
hackers; the latter to revel in the subsequent notoriety such stigmati[z]ation
affords).
TAYLOR, supra note 29, at xiii.
267
See United States v. Prochner, 417 F.3d 54, 60-62 (1st Cir. 2005) (affirming
1382
University of California, Davis
[Vol. 41:1327
yarns about legions of expert hackers to gain new criminal laws,
surveillance powers, and resources.268 The media probably enjoys
higher ratings and ad revenue when reporting on online risks.269
Homeland Security officials specializing in cyberterrorism describe a
world full of evil, renegade hackers, bent on terror and destruction.270
Security vendors do the same.271
The DRM debate is unusual, because the self-interest in using the
Superuser trope appears on both sides. DRM proponents argue that
because they can never win the arms race against powerful users, they
need laws like the DMCA.272 Opponents of DRM argue that the
technology is fundamentally futile because all DRM eventually will be
circumvented.273 Because these partisans and litigants have a vested
interest in building up the Myth of the Superuser, they obscure the
actual reach and influence of Superusers.274
application of enhancement because defendant “hacked” into website and rewrote
“cgi-scripts”); United States v. Lee, 296 F.3d 792, 799 (9th Cir. 2002) (reversing
application of enhancement as applied to designer of fraudulent website); U.S.
SENTENCING GUIDELINES MANUAL § 3B1.3 (2000) (requiring two-level adjustment for
use of special skill).
268
See Elinor Abreu, Net Crime Does Pay For Cops, THE INDUS. STANDARD, Feb. 21,
2000, available at http://www.greenspun.com/bboard/q-and-a-fetch-msg.tcl?msg_id=
002dvE (citing critics who claim that law enforcement inflates or takes advantage of
threat of computer crime to argue for more funding); supra Part II.B.
269
Cf. SUNSTEIN, supra note 163, at 102-03 (citing increased ratings for cable news
stations during reporting about Washington D.C. sniper story of 2002 (citing Johana
Neuman, In a Sniper’s Grip: Media’s Role in Drama Debated, L.A. TIMES, Oct. 24, 2002,
at 16)).
270
NATIONAL STRATEGY, supra note 86, at 6 (stating that “[b]ecause of the
increasing sophistication of computer attack tools, an increasing number of actors are
capable of launching nationally significant assaults against our infrastructures and
cyberspace”).
271
See Levi, supra note 22, at 50 (describing “self-serving PR” of “security
consultants whose income depends on shocking . . . senior executives and government
agencies who complacently fail to spend ‘enough’ money on security”).
272
See Business Software Alliance, Copyright Policy Initiatives to Protect Creative
Works: Digital Millennium Copyright Act (DMCA), http://www.bsa.org/country/
Public%20Policy/Copyright/Copyright%20Policy%20Initiatives%20to%20Protect%20
Creative%20Works.aspx (last visited Apr. 3, 2008) (“The DMCA has helped fuel that
expansion by giving software developers and creative artists the tools they need to go
after pirates who use technology to steal their products and redistribute them en
masse around the world with the click of a button.”).
273
See supra Part II.E.
274
Even those we would not think of as classical advocates or partisans may have a
self-interest in inflating the ability of computer criminals. Consider victims, for
example. Weak network security is often a contributing factor in the success of a
network breach. Victims are unlikely to admit that their security was weak. Low-
2008]
4.
The Myth of the Superuser
1383
The Need for Interdisciplinary Work
Finally, too many experts consider online risk assessment to be
somebody else’s concern. It is not even clear within which field of
study the research should occur. Computer security and computer
science seem the most likely candidates, because these risk
assessments require measuring the rate at which software flaws occur
and understanding whether these flaws are easy or difficult to
exploit.275
Unfortunately, computer security experts often unhelpfully
conclude that all computer software is flawed and that malicious
attackers can and will exploit those flaws if they are sufficiently
motivated.276 The question is not a technology question at all, they
contend, but rather one of means, motive, and opportunity —
questions for criminologists and not engineers.277 Criminologists, for
their part, spend little time studying computer crime, perhaps
assuming that vulnerability-exploit models can be analyzed only by
computer scientists.278
Both sides are right and wrong. Assessing an online risk requires
both computer science and criminology (as well as economics,
psychology, and sociology). Analyses that focus only on some of these
disciplines are short-sighted and often flawed.279
level administrators responsible for security embellish the sophistication of the
attacker to protect their jobs, and their managers do the same thing to minimize
liability or bad publicity. Kevin Poulsen, California Disclosure Law Has National
Reach, SECURITYFOCUS, Jan. 6, 2003, http://www.securityfocus.com/news/1984
(describing “a chronic problem in e-commerce — companies that are hacked are often
reluctant to go public for fear of bad publicity or civil liability”).
275
Wade H. Baker, Loren Paul Rees, & Peter S. Tippett, Necessary Measures:
Metric-Driven Information Security Risk Assessment and Decision Making, 50 COMM.
ACM 101, 102 (2007) (describing information security risk assessment as “the
product of three main factors: frequency of threats/attacks; the likelihood of their
success; and their impact on the organization”).
276
See supra note 214.
277
Email from Author, to Ed Felten, Professor of Computer Science and Policy,
Princeton Univ. (Dec. 19, 2006, 09:50:53 MST) (on file with UC Davis Law Review).
278
For some reason, British criminologists seem much more interested in
computer crime, and many have been cited thus far. See, e.g., TAYLOR, supra note 29;
WALL, supra note 22; Levi, supra note 22.
279
Cf. Levi, supra note 22, at 48 (arguing that problem of computer fraud is “social
and internal” rather than “technological and external”).
1384
University of California, Davis
[Vol. 41:1327
IV. PRESCRIPTIONS FOR MYTH AVOIDANCE
Part III paints a gloomy picture. Fear is a huge problem online.
Because it is abetted by the media, compounded by misunderstandings
about code malleability, and abandoned by experts, dispelling the
Myth of the Superuser is a daunting, difficult task. This Part attempts
to rise to the challenge by proposing steps for dealing with the Myth.
It is important to recognize that this problem and its solutions are not
the same as risk assessment problems encountered elsewhere, such as
environmental regulation.
A. Superuser Harm Is Not Global Warming
In Laws of Fear, Sunstein ultimately prescribes a kind of “expert
cost-benefit analysis” that immunizes policymakers from exaggerated
public fears.280 But because experts have abdicated their role with
online harm, the inputs into his proposed cost-benefit analysis — the
probability of harm and the level of harm caused by the risk — are
very hard to come by.
Sunstein focuses much of his attention on the threat from offsetting
harms, and in particular, the harms caused by regulation.281 For
example, mandating expensive measures to reduce global warming
emissions may raise the cost of consumer goods, which may increase
the levels of poverty.282 Likewise, Part II demonstrated the grave
harms caused by regulations inspired by the Myth of the Superuser.
But despite some similarities, the problems surrounding online harm
and computer crime are different from and prior to the concerns voiced
by Sunstein. Because the malleability of code inspires new stories and
fears as quickly as policymakers can digest them, experts and laypeople
view every Superuser as very likely to cause severe harm. This tilts the
cost-benefit calculation to justify almost any remedial action, such as
increased surveillance powers, harsher penalties, and new restrictions
on conduct. The first task, then, is to move from categorical and
general pronouncements about Superuser harm to specific data about
the likelihood and the severity of online harm.
280
See Dan Kahan et al., Fear of Democracy: A Cultural Evaluation of Sunstein on
Risk, 119 HARV. L. REV. 1071, 1072 (2006).
281
SUNSTEIN, supra note 163, at 23-34.
282
Cf. id. at 31 (describing how Zambian government’s refusal of corn donated
from United States due to concerns about genetically modified food could have led to
35,000 deaths by starvation, according to World Health Organization estimates).
2008]
The Myth of the Superuser
1385
B. Toward a New Empirical Foundation for Online Debates
I propose a new style of discourse for talking about online conflict.
Most importantly, we must demand more extensive and improved
quantitative and qualitative empirical evidence. Before we can do that,
we need to address the pervasive misuses of anecdote. Storytelling will
always be part of our debates, but we must become better producers
and more discriminate consumers of the stories that are told.
1.
A Moratorium on Suspect Rhetoric
I call for a moratorium on urban myths and just-so stories — tall
tales in which the power of the attacks and evasiveness of the attackers
increase with each retelling. The earlier discussion of the 1996 Report
highlights some of the harmful and pervasive rhetorical tools that are
too often used and never challenged.283 We should ignore, or at least
give very little weight to, stories built upon any of the following
devices.
a.
Mere Possibilities
Logicians describe the “appeal to probability” as the logical fallacy
that results when mere possibility is confused with likelihood.284 Just
because something may happen, people conclude that it will happen
and that it deserves a response. This is especially troublesome online
because the generativity of software seems to makes everything
possible.285
Bare arguments about what may or could occur should be given
almost no weight by policymakers. Given the limits of human
bandwidth, most undesirable possibilities will never occur online.
Nobody has ever been killed as the result of an online attack.286 The
Internet has never “crashed”287 and never will.288 Further, I could find
283
See supra note 19.
See Levi, supra note 22, at 50 (discussing official sources that “conflate
experience of [risk] with theoretical risk from computer crime” (emphasis omitted)).
285
See Zittrain, supra note 156 passim; supra Part III.B.
286
See Krebs, supra note 85.
287
I am unconvinced that the Internet has ever crashed, although, of course, much
turns on how I define “crash.” In the 2000 Denial of Service attacks that struck
Yahoo!, eBay, and others, the vast majority of Internet sites and Internet users
continued to operate with little to no noticeable effect. Similarly, although the Morris
Worm had a significant effect on a much-smaller Internet in 1988, service was
restored in a matter of hours, estimates suggest that only 10% of hosts were infected,
and it is almost inconceivable that we would see an outage of similar scale today. See
284
1386
University of California, Davis
[Vol. 41:1327
no documented cases of criminals using software booby-traps to
destroy evidence.
b.
Secondhand Information
Policymakers should be skeptical of the storyteller who is many
steps removed from the technical expert or firsthand witness. As
children learn from the game of telephone, stories lose important
details with each step in the chain. Thus, the longer the chain, the
more the story should be doubted.
Furthermore, because the devil is in the technical details, stories of
Superuser power should be second-guessed when the speaker lacks
the technical knowledge to be trusted with getting the details right.
Similarly, stories should be doubted when they have crossed from
participant to advocate. All of these problems arise, for example,
when stories about powerful computer criminals are reported by
legislative affairs officers pressing changes in the laws.289
c.
Secrecy
Moreover, too many Superuser storytellers enshroud the most
important details within veils of secrecy. Security consultants claim
the need to protect their clients’ identities.290 Prosecutors refuse to
reveal the details of ongoing investigations.291 The intelligence
community refuses to talk about anything.292
Paul Graham, The Submarine, http://www.paulgraham.com/submarine.html (last
visited Apr. 3, 2008).
288
See supra note 87.
289
See Declan McCullagh, Porn Spammers to Face Jail in U.S., ZDNET.CO.UK, July 9,
2003,
http://news.zdnet.co.uk/internet/0,1000000097,2137288,00.htm
(quoting
Assistant Attorney General William Moschella on Department of Justice’s positions on
then-pending antispam law); Press Release, U.S. Dep’t of Justice, Justice Department
Announces William E. Moschella as New Principal Associate Deputy Attorney General
(Oct. 2, 2006), available at http://www.usdoj.gov/opa/pr/2006/October/06_odag_
666.html (reporting Moschella was in charge of DOJ’s Office of Legislative Affairs in
July 2003).
290
See supra note 253 and accompanying text.
291
See, e.g., Christopher Hayes, But Can He Hack Prison?, CHI. READER, Aug. 19,
2005, available at http://www.chicagoreader.com/pdf/050819/050819_cover.pdf.
(describing investigation into hack of website and reporting that “[a]n FBI
spokesperson says the Bureau won’t comment on an ongoing investigation”).
292
See Editorial, The Dangerous Comfort of Secrecy, N.Y. TIMES, July 12, 2005, at
A20 (“The Bush [A]dministration is classifying the documents to be kept from public
scrutiny at the rate of 125 a minute.”).
2008]
The Myth of the Superuser
1387
In particular, the need to protect victims’ rights is a reason why
secrecy is given too much deference. For example, corporate victims
will often not reveal that their security has been breached.293 Although
such corporations usually claim to worry about customer privacy, they
more likely worry about plummeting customer confidence.294
While many of these obligations of secrecy are no doubt valid and
important, those who labor beneath these restrictions should not be
allowed to spin half-stories to influence policy without being required
to pierce the veil in return. Stories lacking details, even for legitimate
reasons, should not drive policy. I argue that most of the time, they
should be given no weight whatsoever.
Also, legislatures should consider laws to eliminate unnecessary
secrecy. For example, California now requires companies to publicly
disclose security breaches in computer systems that house personal
information.295 Although many companies have criticized this law and
opposed similar legislation in other states and in Congress, many
privacy and security experts praise the law.296 Not only does it keep
the public better informed about the uses of personal information, but
some argue that it provides better data for researchers who want to
weigh online risks.297
As security researcher Adam Shostack
explained:
293
See Richard P. Salgado, U.S. Dep’t of Justice, Working with Victims of Computer
Network Hacks, USA BULL., Mar. 2001, available at http://www.usdoj.gov/
criminal/cybercrime/usamarch2001_6.htm (“Intrusion victims, however, are often
even more reluctant to call law enforcement than other business victims.”).
294
Id. (reporting that some industry participants in DOJ working group say they
did not report past hacks to law enforcement because “the fact of the intrusion will
become public knowledge, irreparably shaking investor confidence and driving
current and potential customers to competitors who elect not to report intrusions”).
295
See CAL. CIV. CODE §§ 1798.29, 1798.82 (West 2002). This is popularly
referred to as SB 1386.
296
See Jaikumar Vijayan, Breach Notification Laws: When Should Companies Tell
All?, COMPUTERWORLD, Mar. 2, 2006, available at http://www.computerworld.com/
securitytopics/security/story/0,10801,109161,00.html (citing “growing industry
consensus that security breach notification laws have forced companies to take more
responsibility for the data they own”).
297
Posting of Adam Shostack to Emergent Chaos, http://www.emergentchaos.com/
archives/2007/03/security_breaches_are_goo.html (Mar. 29, 2007, 00:43) (describing
presentation with slides that can be found at Adam Shostack, Security Breaches Are
Good for You (2007), http://www.homeport.org/~adam/Security%20Breaches%
20are%20good%20for%20you.pdf). The study of newspaper accounts of security
breaches by Erickson and Howard described supra note 57, noted a significant
increase in reporting since the passage of SB 1386.
1388
University of California, Davis
[Vol. 41:1327
The reason that breaches are so important is that they provide
us with an objective and hard to manipulate data set which we
can use to look at the world. It’s a basis for evidence in
computer security. Breaches offer a unique and new
opportunity to study what really goes wrong. They allow us to
move beyond purely qualitative arguments about how bad
things are, or why they are bad, and add quantifatication
[sic].298
d.
Undetectable Power
Because some Superusers’ tools are so hard to detect, we will not
even know when they have been used. Thus, there is no value
pointing to them in policy debates. Consider steganography. A close
relative of encryption, steganography involves hiding things in plain
view.299 People use steganographic software to encode messages or
files within other files.300 For example, text messages can be hidden
within image files that can then be placed on public websites,
remaining hidden in plain view.301 Although researchers have
developed tools to detect some forms of steganography, the research is
difficult to conduct and unlikely to be very good at detecting new
forms of steganography.302 Consequently, the spread of steganography
is nearly impossible to count or otherwise profile.
The empirical difficulty at the heart of the Myth of the Superuser is
at its worst with secret, undetectable tools such as this. Nevertheless,
steganography is often cited by scholars trying to (1) justify giving the
NSA or FBI more invasive surveillance authority by stating that
cunning terrorists are capable of using advanced Internet
technology,303 or (2) prove that new surveillance powers are futile
298
Shostack, supra note 297.
See NIELS PROVOS & PETER HONEYMAN, CTR. FOR INFO. TECH. INTEGRATION,
STEGANOGRAPHIC
CONTENT
ON
THE
INTERNET
3
(2001),
DETECTING
http://www.citi.umich.edu/techreports/reports/citi-tr-01-11.pdf.
300
Id.
301
Id.
302
Cf. id. (reporting that scan of two million images on eBay had failed to identify
any steganographic messages).
303
See Orin Kerr, Internet Surveillance Law after the USA PATRIOT Act: The Big
Brother that Isn’t, 97 NW. U. L. REV. 607, 607 (2003) (arguing that amending Internet
surveillance laws would help War on Terror because terrorists were known to use
advanced Internet technologies); Michael J. Woods, Counterintelligence and Access to
Transactional Records: A Practical History of USA PATRIOT Act Section 215, 1 J. NAT’L
SECURITY L. & POL’Y 37, 37 (2005) (explaining that former chief of FBI National
299
2008]
The Myth of the Superuser
1389
because criminals will simply turn to more secretive ways to
communicate.304 These arguments are supported by journalists who
have written articles about how al Qaeda or Osama bin Laden might
be using steganography.305 Thus, because claims about the possible
use of Superuser tools like steganography are speculative and
inherently irrefutable, they should never be considered effective
support in policy debates.
2.
Improving the Empirical Foundation: Toward Better FactFinding
Ignoring suspect anecdotes will set up the conditions for better
policymaking, but this is merely ground-clearing. To truly assess the
power of the Superuser, we need better, more reliable, more
persuasive facts, and we need to better use the facts we have. Our goal
should be to discover whether Superusers or ordinary users account
for more harm.
In the course of writing this Article, I had an enlightening email
exchange with a technologist at the Electronic Frontier Foundation.306
He took exception to my hostility to the idea that all DRM is
inherently flawed.307 Early in our exchange, he pointed to the Darknet
paper as part of his proof of this hypothesis.308 Although he had not
changed my mind (and I do not think I had changed his), by the end
of the exchange, instead of relying on a single reference to the Darknet
Security Law Unit argued that transactional record information is valuable when
hunting terrorists because content information can be obscured, for example, with
steganography).
304
See Caspar Bowden, Closed Circuit Television for Inside Your Head: Blanket
Traffic Data Retention and the Emergency Anti-Terrorism Legislation, 2002 DUKE L. &
TECH. REV. 5, 5 (2002) (arguing against part of then-proposed UK Anti-Terrorism
Crime and Security Bill because undetectable communication via steganography
would remain undetected).
305
See Jack Kelley, Terror Groups Hide Behind Web Encryption, USA TODAY, Feb. 5,
2001, at 7A (citing unnamed “U.S. and foreign officials” for proposition that bin
Laden is using steganography); Kevin Maney, Osama’s Messages Could Be Hiding in
Plain Sight, USA TODAY, Dec. 19, 2001, at B6 (acknowledging that “no actual evidence
has been found of al[-]Qaeda using” steganography, but engaging in hype
nevertheless). The author of the first article, Jack Kelley, was revealed to have made
up many of his stories and sources, although the ensuing review of his work did not
specifically highlight this article. See Blake Morrison, Ex-USA Today Reporter Faked
Major Stories, USA TODAY, Mar. 19, 2004, at 1A.
306
See Email from Derek Slater, Activism Coordinator, Electronic Frontier Found.,
to Author (Aug. 15, 2006, 08:50:05 PST) (on file with UC Davis Law Review).
307
See id.
308
See id.
1390
University of California, Davis
[Vol. 41:1327
paper, he provided paragraphs of history about the evolution of DRM
and DRM crackers and analyses about why DRM is doomed to
failure.309 It occurred to me that this exchange exemplified the result
that I seek. Sweeping away the rhetorical shortcut that the Myth of
the Superuser represents brings us closer to understanding the true
nature of the problems we are trying to solve.
a.
Better Using Facts
Before we discover a single new fact, we can make better use of the
facts that we have. Even with the moratorium on suspect rhetoric, we
cannot abandon anecdotes completely. Because it will be a long time
before we gather enough reliable statistics to fully illuminate the risks
we face, part of what we must consider will come from stories.
Recall, however, that we are gullible, passive consumers of
Superuser anecdotes. To improve, every time we are faced with a
Superuser anecdote we should look for competing narratives to help
us decide whether the Superuser power or the ordinary user story is
the accurate one. For example, despite all of the focus on evil, “blackhat” Superusers, not enough focus is on their counterparts, the “whitehat” Superusers who develop countermeasures. Virus writers are
opposed by the antivirus community, spammers battle spam filterers,
and system crackers clash with system securers.310 Policymakers too
rarely assess how well white-hat technologists are solving problems on
their own. In some of these arms races, the countermeasure
community may hold the upper hand, and if we learn to recognize the
hallmarks of this arms race, we can learn to wait rather than
regulate.311
Consider virus writing. The uninformed observer would accept the
Myth that powerful, malevolent Superusers create and release
damaging viruses and worms and that they do so with impunity. This
observer might urge policymakers to expand crimes that punish virus
writers. A more informed observer would take into account the
following successes of the antivirus community.312 First, up-to-date
309
See id.
See, e.g., Andrew D. Smith, McAfee Improves Computer Security, but It's Not
Perfect, DALLAS MORNING NEWS, Oct. 31, 2007, at 1D (interviewing CEO of computer
security software firm McAfee).
311
See Ohm, supra note 234, at 1984-85 (describing countermeasures designed by
Usenet participants and arguing that regulators should consider them when deciding
whether to regulate).
312
Felten’s testimony to a House subcommittee inspired this analysis. See Piracy of
Intellectual Property on Peer-to-Peer Networks: Hearing Before the Subcomm. on Courts,
310
2008]
The Myth of the Superuser
1391
antivirus software does a pretty good job of preventing the spread of
viruses written in the past. Second, computer users who diligently
install OS patches and software updates are relatively immune to
newly created viruses. Third, only a small percentage of new viruses,
particularly those written to exploit new vulnerabilities, will infect the
computers of these diligent users. Fourth, only Superusers with
significant computer programming ability, plenty of spare time, and
access to a community of like-minded attackers, will succeed in
infecting these machines; the rest of the would-be virus writers —
those with less ability, time, and community — will write duds.
On the other hand, this conclusion perhaps ignores the most
relevant observation: many people do not update their virus software
and do not diligently install OS patches.313 Non-Superuser attackers
using old tools, some of which are packaged for script kiddies with
easy-to-use graphical interfaces, can successfully infect these victims’
computers.314 This goes back to the idea that with viruses, even
ordinary users can do great damage.
b.
Using Better Facts
Scholars tend to be pessimistic about finding meaningful statistics to
measure the basic occurrence of online harm.315 Perhaps they have
not been looking in the right places or asking the right people. There
are several promising, rarely tapped groups of researchers attempting
to count computer vulnerabilities and exploits.
First, an entire industry for conducting risk assessments of online
harm has arisen. Risk assessment theory spans a wide variety of
the Internet, and Intellectual Property of the H. Comm. on the Judiciary, 107th Cong., at
231 (2002) (testimony of Edward W. Felten, Associate Professor of Computer Science,
Princeton University), available at http://commdocs.house.gov/committees/judiciary/
hju81896.000/hju81896_0.HTM. “[A]nalysis of the arms race between virus writers
and antivirus companies leads to the prediction that antivirus products will be able to
cope almost perfectly with known virus strains but will be largely helpless against
novel viruses. This is indeed what we observe.” Id.
313
See David Talbot, The Internet Is Broken, TECH. REV., Dec. 20, 2005,
http://www.technologyreview.com/InfoTech/wtr_16055,258,p1.html
(“[D]ifferent
people use different patches and not everyone updates them religiously; some people
don’t have any installed.”).
314
See Matt Hines, Unpatched Machines ‘Net’s Biggest Threat,’ ZDNET.CO.UK, Apr.
26, 2005, http://news.zdnet.co.uk/security/0,1000000189,39196317,00.htm.
315
See Susan Brenner, Cybercrime Metrics: Old Wine, New Bottles?, 9 VA. J.L. &
TECH. 13, *1, n.3 (2004) (noting that there are no “measures and benchmarks for the
incidence and damage caused by” computer crime).
1392
University of California, Davis
[Vol. 41:1327
quantitative and qualitative methods.316 The insurance industry, for
example, now sells insurance policies relating to computer security.317
Their studies and the methodologies they employ are rarely cited by
policymakers in debates over online threats. This is partly because the
industry does not produce the type of detailed analysis that would be
most useful for policymakers. Perhaps others can translate the
industry’s reports to tease out details to make them more useful for
policymakers.
A related development is the burgeoning economics of information
security discipline.318 These researchers look at computer and
network security as more than just the study of software
vulnerabilities, exploits, and countermeasures. They try to account for
incentives, externalities, and markets that lead to insecurity.319
There are several other pools of untapped statistics about online
crime. For example, many organizations monitor online harms for
different reasons. Companies that sell virus and spyware scanning
software keep statistics about malicious code activity.320
The
Recording Industry Association of America and the Motion Picture
Association of America monitor peer-to-peer networks with advanced
data-collection “spiders” to track the distribution of their copyrighted
works on those networks.321 More disinterested noncommercial
entities such as the SANS Internet Storm Center collect information
about threats on the Internet.322 For example, the Honeynet Project’s
316
See CHARLES A. SHONIREGUN, IMPACTS AND RISK ASSESSMENT OF TECHNOLOGY FOR
INTERNET SECURITY 68-79 (2005) (describing four quantitative and two qualitative
approaches for conducting risk assessments and concluding that none model risk from
“technology enhanced information” well).
317
But see Ross Anderson & Tyler Moore, The Economics of Information Security,
SCI., Oct. 27 2006, at 613 (noting that “the cyber-insurance market is both
underdeveloped and underused”).
318
See generally id. (providing primer for new economics of information security
discipline).
319
Id.
320
See McAfee Avert Labs, Technical White Papers, http://www.mcafee.com/
us/threat_center/white_paper.html (last visited Apr. 3, 2008); Sophos, White Papers,
http://www.sophos.com/security/whitepapers/ (last visited Apr. 3, 2008); Symantec
Security Response, White Papers, http://www.symantec.com/enterprise/security_
response/whitepapers.jsp (last visited Apr. 3, 2008).
321
See Benny Evangelista, Firm Sleuths Out Illegal File Sharers: BayTSP Tracks
Down IP Addresses, IDs of Music Downloads, S.F. CHRON., July 21, 2003, at E1
(describing firm that markets monitoring services of peer-to-peer networks to
recording industry copyright owners).
322
SANS Institute, SANS Internet Storm Center, http://isc.sans.org/ (last visited
Apr. 3, 2008).
2008]
The Myth of the Superuser
1393
volunteers set up purposefully vulnerable computers on the Internet
to monitor and study computer intrusions.323 Policymakers and
scholars should more aggressively tap into these sources to better
profile online threats.
There are no doubt other sources of information and experts in
other fields who could contribute to an effort to measure online harm
accurately. We should embrace contributions from all of these
sources. But bearing in mind the problems of expertise described in
Part III.C, we should search for those without vested interests. For
example, we should look for those who specialize in fields across an
interdisciplinary spectrum with backgrounds in statistics and
computer technology.
c.
Discovering New Facts
Finally, policymakers should commission new studies to measure
the actual incidence of Superuser threats. A good example on the
horizon is the ongoing “National Computer Security Survey”
cosponsored by the DOJ’s Bureau of Justice Statistics and the
Department of Homeland Security’s National Cyber Security Division
and administered by the RAND Corporation.324 Unlike the CSI/FBI
Survey, the latest effort is an ambitious attempt to use rigorous
methods to obtain statistically meaningful numbers.325 According to
the survey group, this “is the first and only survey to provide official
national statistics on the extent and consequences of computer
security incidents within businesses across all industry sectors.”326
Along with other promising methodological choices, the researchers
are canvassing businesses in thirty-seven industry sectors, and they
plan to repeat the survey every one to two years.327
C. The Anti-Precautionary Principle
In Laws of Fear, Sunstein critiques the Precautionary Principle,
which holds that when faced with evidence of a harm, but uncertainty
323
The Honeynet Project, Honeynet Project News, http://www.honeynet.org/ (last
visited Apr. 3, 2008).
324
See RAND Corp., DOJ/DHS National Computer Security Survey,
http://www.ncss.rand.org (last visited Apr. 3, 2008) [hereinafter RAND study
website].
325
For a description of the CSI/FBI Survey and a summary of its criticisms, see
supra notes 66-74.
326
RAND study website, supra note 324.
327
RAND study website, supra note 324.
1394
University of California, Davis
[Vol. 41:1327
about the harm’s probability, regulators should act to prevent the
harm.328 Given the historically impoverished empirical evidence that
has led to the Myth of the Superuser and the litany of harms that arise
when policymakers respond to the Myth, I propose a default — the
Anti-Precautionary Principle. In any online conflict, the presumption
should be to regulate only the ordinary user unless facts suggest that
the Superuser is a significant threat.
Of course, especially given the Myth, there will often be the
suggestion that some Superusers exert power. The Anti-Precautionary
Principle’s presumption will not be invoked. What then? First, if
reliable facts establish there are very few Superusers with little
aggregate impact, the policymaker should obey the Anti-Precautionary
Principle and act as if the Superusers do not exist. That is, when a
conflict involves ordinary users in the main and Superusers only at the
margins, the harms resulting from regulating the few cannot be
justified.
What if the facts point to Superuser domination and significant
potential harm? It depends on the strength of that evidence. If the
evidence is sound and convincing, the Anti-Precautionary Principle
should likely give way, and policymakers should consider regulating
the Superuser mindful of the possible harms discussed in Part II.329
When the evidence is contingent or weak, a pure balancing should
apply, comparing the harm caused by the Superusers with the harms
caused by regulating.
As an important exception, the Anti-Precautionary Principle should
probably yield in the face of horrific potential harms. Policymakers
should defer to law enforcement in the face of plausible stories — for
example, about terrorist attacks on computer networks that could
result in deaths — supported by at least anecdotal evidence that relies
on more than the rhetorical devices described in Part IV.B.1.330 As I
328
See SUNSTEIN, supra note 163, at 18-20. Sunstein notes that the Precautionary
Principle embodies a range of 20 or more definitions of varying levels of strength. Id.
at 18. Weak forms of the Principle (noting “lack of decisive evidence of harm should
not be a ground for refusing to regulate”) are unobjectionable to Sunstein, but it is the
strongest forms that he criticizes. Id. The form cited in the text is closer to Sunstein’s
strong forms, and will be the definition used in this Article. Much of what will be said
will apply to many different possible definitions.
329
See infra Part IV.D (discussing how to craft minimally harmful laws in this
situation).
330
Sunstein talks about a similar Anti-Catastrophe Principle. SUNSTEIN, supra note
163, at 109-15. In the face of an uncertain probability of catastrophe, Sunstein
concedes, a form of the Precautionary Principle should apply. Id.
2008]
The Myth of the Superuser
1395
suggested earlier, however, experts seem to agree that such attacks are
not likely.331
To illustrate my proposal, consider the DOJ’s recently proposed
amendment to § 1030. Under current federal law, causing damage to
a computer is a crime only if it causes a sufficient amount or specific
type of loss, which for most cases means the victims must have
suffered more than $5000 in aggregate losses.332 There may be fear
that the $5000 loss limit will hinder the prosecution of people who
deploy and use what are known as “botnets.” Whereas traditional
computer intruders cause damage or rifle through private files on
computers one at a time, a botnet operator collects computers by the
hundreds or thousands, causing very little damage to any individual
computer.333 For example, he assembles his “zombie computer army”
for future use, ordering them all someday to attack a targeted web
In this situation, during the “quiet
server simultaneously.334
collection” phase of the botnet and before the ultimate attack, it would
be difficult to show that aggregate harms total $5000.
At a “brown bag lunch hearing” on Capitol Hill,335 speakers
proposed reducing the $5000 threshold.336 In considering whether to
turn this proposal into legislation, Congress should analyze the “Myth
of the Botnet General” under my rubric. First, because there have
been some confirmed botnet cases, the extent and veracity of this
evidence must be scrutinized.337 The overheated rhetoric will
probably suggest that botnets are enormous problems that can cause
significant harms, but the evidence made public so far seems
predictably unrigorous and largely anecdotal.338 Ideally, better, more
rigorous statistics — perhaps from virus and spyware company
studies — could establish the true nature of the risk.
331
See supra note 85.
18 U.S.C. § 1030(a)(5)(B)(i) (2000).
333
See John Markoff, Attack of the Zombie Computers Is Growing Threat, N.Y. TIMES,
Jan. 7, 2007, at A1; Press Release, U.S. Dep’t of Justice, Computer Virus Broker
Arrested for Selling Armies of Infected Computers to Hackers and Spammers (Nov. 3,
2005), http://www.usdoj.gov/criminal/cybercrime/anchetaArrest.htm.
334
Markoff, supra note 333.
335
Association for Computing Machinery, USACM Technology Policy Weblog,
Briefing:
Learning about the Threats from Botnets (Apr. 20, 2007),
http://usacm.acm.org/weblog/index.php?p=490.
336
See Posting of Ed Felten to Freedom to Tinker Blog, http://www.freedom-totinker.com/?p=1150 (Apr. 26, 2007, 10:41) (reporting on proposals made by others
during hearing).
337
Markoff, supra note 333.
338
See id.
332
1396
University of California, Davis
[Vol. 41:1327
These risks must be balanced against the risks created by getting rid
of the $5000 threshold. The threshold serves an important check: it
minimizes trivial prosecutions. Many annoying acts occur on
networks every day. Spam filters delete nonspam; practical jokes are
played in offices on coworkers; files are accidentally deleted from
shared servers. Under current law, these non-Superuser acts are not
usually prosecuted even though they may fall within the broad and
vague conduct elements of the statute’s damage prohibitions.339
Prosecutors and agents who receive a call from a “victim” of one of
these acts will almost certainly decline to prosecute or even investigate
because they know they will never meet the $5000 loss threshold.
If Congress were to remove the threshold, law enforcement agents
could subject harmless people like the office prankster to invasive
search and surveillance, and overzealous prosecutors could then bring
charges. Given the weak empirical proof about the threat posed by
botnets, the risks from removing the threshold outweigh the risks that
a botnet owner will someday be apprehended but not prosecutable.340
D. Two Additional Difficulties
1.
Regulating Actual Superusers
Even if Congress adopts the Anti-Precautionary Principle and begins
to demand better empirical evidence, it may conclude that the
Superuser threat outweighs the harm from regulating. I am not
arguing that Superusers should never be regulated or pursued. But
given the checkered history of the search for Superusers — the
overbroad laws that have ensnared non-Superuser innocents; the
amount of money, time, and effort that could have been used to find
many more non-Superuser criminals; and the spotty record of law
enforcement successes — the hunt for the Superuser should be
narrowed and restricted. Policymakers seeking to regulate the
339
For example, § 1030(a)(5)(A)(iii) criminalizes access in excess of authorization
that causes damage (over $5000) and the subsection has no mens rea requirement. 18
U.S.C. § 1030(a)(5)(A)(iii) (2000). Even unintentional, nonreckless, nonnegligent
damage may violate this provision, a misdemeanor for first-time offenders. Id.
340
Given the breadth of substantive federal criminal law, there are almost certainly
other federal crimes that could be used to prosecute the Botnet General. Cf. Peter J.
Henning, Maybe It Should Just Be Called Federal Fraud: The Changing Nature of the
Mail Fraud Statute, 36 B.C. L. REV. 435, 437 (1995) (“The appeal of the mail fraud
statute is its malleability: federal prosecutors can pursue investigations with the
knowledge that, in bringing an indictment, they will not be hampered by technical
jurisdictional restrictions often found in other federal criminal statutes.”).
2008]
The Myth of the Superuser
1397
Superuser can adopt a few strategies to narrowly target Superusers and
minimally impact ordinary users.
The chief evil of past efforts to regulate the Superuser has been the
inexorable broadening of laws to cover metaphor-busting, impossibleto- predict future acts. To avoid the overbreadth trap, legislators
should instead extend elements narrowly, focusing on that which
separates the Superuser from the rest of us: his power over technology.
They should, for example, write tightly constrained new elements that
single out the use of power, or even, the use of unusual power.
Consider, for example, the Superuser-induced phrase, “access
without authorization,” an element of several different computer
crimes.341 Professor Orin Kerr has noted that courts have construed
this vague phrase to apply to many ordinary acts that do not seem to
amount to computer hacking, and in many cases, seem unworthy of
criminal or even civil sanction.342 Recall Explorica, where a travel
agency was found to have engaged in unauthorized access when it
scraped information from its competitor’s public website, even though
it could have used a web browser to obtain the exact same
information.343 In another example, a court held that a man accessed
his employer’s computer without authorization, which he ordinarily
had full permission to access, because he, a faithless employee, sent
files to a competitor for whom he planned to work in the near future.344
Unhappy with the breadth of these results, Kerr proposed a
Superuser-centric amendment. He argued that an act is not done
“without authorization” under § 1030 unless the actor circumvented
“code-based restrictions on computer privileges.”345 This formulation
creates two requirements: first, the computer accessed must have had
some sort of “code-based” (i.e., software or hardware based) security
or other “restriction[] on computer privileges,” and second, the actor
had to “circumvent” that restriction.346 Under this framework, visits
341
See 18 U.S.C. § 1030(a)(1)-(6).
Kerr, supra note 103, at 1649-50 (arguing for narrower interpretation of
“without authorization” in statute and criticizing cases that have read phrase broadly
for “sacrificing a great deal of freedom for a small (and arguably minimal) gain in
privacy and security”).
343
EF Cultural Travel BV v. Explorica, Inc., 274 F.3d 577, 583 (1st Cir. 2001).
Granted, scraping the site with a mere web browser would have been very labor
intensive. Cf. id. (calling manual alternative “theoretically . . . possible”).
344
Int’l Airport Ctrs., L.L.C. v. Citrin, 440 F.3d 418, 419, 421 (7th Cir. 2006).
345
Kerr, supra note 103, at 1656.
346
Id.
342
1398
University of California, Davis
[Vol. 41:1327
to public websites would probably not suffice, and breaking an
employment contract certainly would not.
As another example, consider again the DMCA, which prohibits the
DRM circumvention.347 One reason the law has been criticized since
before its passage is that it places no serious limits on how
sophisticated a lock must be before it gains the backing of the
prohibition.348
Although the law extends only to DRM that
“effectively controls access” to a copyright-protected work, that phrase
is defined to mean that the DRM “in the ordinary course of its
operation, requires the application of information, or a process or a
treatment, with the authority of the copyright owner, to gain access to
the work.”349 Courts have interpreted this to place almost no
restrictions on the level of sophistication required — even DRM that is
trivial to unscramble will satisfy the low hurdle for protection.350
The DMCA can be rewritten to focus more specifically on the
Superuser while not casting its overbroad net on ordinary users. For
example, “effectively controls access” could be amended to apply only
to digital locks that pass a particular threshold of complexity. Perhaps
this could be defined in terms of encryption algorithms that have been
sufficiently peer-reviewed or by uses of a 128-bit symmetric key
length.351 Perhaps a regulatory process can define the level of
technology protected.
The point is to try to create a balance between addressing the
harm — indiscriminate cracking of DRM and rampant copyright
infringement — and ensuring that average, ordinary users are not
prosecuted for doing ordinary things or investigated for months before
347
17 U.S.C. § 1201(a)(1)(A) (2000).
See Cohen, supra note 34, at 172-73.
349
17 U.S.C. § 1201(a)(3)(B).
350
Universal City Studios, Inc. v. Reimerdes, 111 F. Supp. 2d 294, 318 (S.D.N.Y.
2000) (“Finally, the interpretation of the phrase ‘effectively controls access’ offered by
defendants at trial-viz., that the use of the word ‘effectively’ means that the statute
protects only successful or efficacious technological means of controlling access —
would gut the statute if it were adopted.”).
351
Holding everything else constant, the longer (the more bits) the key used, the
harder it is (the more computation time it requires) to crack encrypted ciphertext. See
BRUCE SCHNEIER, APPLIED CRYPTOGRAPHY 152 (2d ed. 1996) (“Two parameters
determine the speed of a brute-force attack: the number of keys to be tested and the
speed of each test.”). Although the details are well outside the scope of this Article,
cryptographic algorithms come in two broad flavors, public-key and symmetric-key
cryptography. For any given key length, it is easier to crack a public key than a
symmetric key. Id. at 166 tbl.7.9. A 128-bit key length is still somewhat respectable
for a symmetric key algorithm, but it is laughable for public key cryptography. Id.
348
2008]
The Myth of the Superuser
1399
the pall of suspicion passes over them.352 In short, the idea is to craft
laws that are tailored to the type of power that Superusers wield.353
2.
Regulating Prometheus: The Problem of Script Kiddies
Finally, sometimes Superusers empower ordinary users with easyto-use software.354 These “script kiddies,” as they are known in
computer security jargon, are like modern Prometheuses given great
power from above. How does this threat fit within a model that urges
inattention to Superusers?
First, the script kiddie should be part of the cost-benefit accounting.
When policymakers are balancing competing harms, they should
factor in the possibility that Superusers will empower legions of
ordinary users.
As always, bare assertions that this kind of
empowerment is likely or inevitable should be met with suspicion.
After all, there is probably a “Myth of the Script Kiddie.”
Second, and more controversially, there are steps to keep Superusers
and script kiddies apart. Professor Randy Picker has proposed what
he terms an “incentive wedge” to keep honest, ordinary users from
using DRM circumvention tools.355
Specifically, he proposes
embedding digital music and movies with personally identifiable
information that can direct future investigators back to the source.356
352
See Electronic Frontier Foundation, US v. ElcomSoft & Sklyarov FAQ (Feb. 19,
2002),
http://www.eff.org/IP/DMCA/US_v_Elcomsoft/us_v_sklyarov_faq.html
(describing indictment of Dmitri Sklyarov for circumventing some very weak
protection schemes of DRM on Adobe eBook reader).
353
The obvious downside to this proposal is that defining criminal acts with
respect to technical power can lead to the guilt by association problem described
above. If lawmakers create prohibitions defined by a person’s technical sophistication
and power, other elements of those prohibitions must protect researchers, students,
security professionals, and others, who act powerfully but without evil intent or harm.
For example, the harm elements of the prohibition should be definite and clear, so
that a researcher who circumvents DRM but does not create downstream copies or
release automated tools will not be covered.
354
See Erickson & Howard, supra note 57, at 12 (“Since knowledge and tools
developed by more experienced hackers can easily be obtained on the Internet, the
capability to penetrate insecure networks has propagated outside of the legitimate
hacker community to other groups, ranging from inexperienced teenagers to
international crime syndicates.”).
355
Picker, supra note 33, at 49.
356
Id. at 69. Ed Felten disagrees. Posting of Ed Felten to Freedom to Tinker Blog,
http://www.freedom-to-tinker.com/?p=980 (Feb. 22, 2006, 13:54) (calling Picker’s
idea “an instructive idea, but not a practical one”). Apple Computer began embedding
the names of customers into the songs they purchased through iTunes. May Wong,
Questions Raised over iTunes User Data, USA TODAY, June 5, 2007, available at
1400
University of California, Davis
[Vol. 41:1327
Even more controversially, faced with a conflict in which Superusers
have not yet empowered script kiddies, perhaps Congress should ban
the creation or distribution of script kiddie empowering tools. This
idea gives me significant pause because of Congress’s checkered track
record at drafting this kind of regulation, but some lessons can be
learned by looking at past attempts. The DMCA, which prohibits the
creation and distribution of DRM circumvention tools,357 and § 2512
of the Wiretap Act, which prohibits the creation and distribution of
particular types of wiretapping devices,358 are two prominent examples
of laws that prohibit the distribution of software. Comparing these
two laws raises an interesting question. Why is the DMCA’s
distribution prohibition so controversial while § 2512 is not? The
answer may serve as a map for creating less controversial software
distribution bans.
There are some obvious possibilities. First, § 2512 is very rarely
prosecuted.359 Although the DMCA also rarely leads to criminal
charges, it is often rattled like a saber by civil litigants.360 Second, §
2512 predated the spread of the Internet and the rise of online civil
liberties groups.361 If § 2512 were proposed anew today, it would
almost certainly meet fierce opposition.
There is a less intuitive, more intriguing possibility: the DMCA
targets technology that has potentially beneficial, legal uses. Many
law-abiding citizens would like to circumvent DRM to backup their
http://www.usatoday.com/tech/products/services/2007-06-05-itunes-drm-free_N.htm.
357
17 U.S.C. § 1201(a)(2) (2000).
358
18 U.S.C. § 2512 (2000).
359
A search of Westlaw’s ALLFEDS database for the phrase “18 U.S.C. § 2512” on
February 6, 2008 returned 19 cases. Almost all of these cases were civil lawsuits
brought by DirectTV seeking to use the provision against people selling and buying
satellite TV descrambler chips. See, e.g., DirectTV, Inc. v. Barnes, 302 F. Supp. 2d 774
(W.D. Mich. 2004). The search returned four criminal prosecutions. United States v.
Biro, 143 F.3d 1421, 1422 (11th Cir. 1998); United States v. Daniels, 978 F.2d 415,
416 (8th Cir. 1992); United States v. Novel, 444 F.2d 114, 114 (9th Cir. 1971); United
States v. Spy Factory, Inc., 951 F. Supp. 450, 451-52 (S.D.N.Y. 1997).
360
See Chilling Effects Clearinghouse, http://www.chillingeffects.org (last visited
Apr. 3, 2008) (collecting and displaying copyright-related cease and desist letters).
361
A form of § 2512 was part of the original Wiretap Act enacted in 1968.
Omnibus Crime Control and Safe Streets Act of 1968, Pub. L. No. 90-351, §802, 82
Stat. 214. In comparison, the EFF was founded in 1990, EFF, About EFF,
http://www.eff.org/about (last visited Apr. 3); the Center for Democracy and
Technology in 1994, CDT, Summary of CDT Activities 2000 — Work Plan 2001,
http://www.cdt.org/mission/activities2001.shtml (last visited Apr. 3, 2008); and the
Electronic Privacy Information Center also in 1994, EPIC, About Epic,
http://epic.org/epic/about.html (last visited Apr. 3, 2008).
2008]
The Myth of the Superuser
1401
movies and software, time-and-space shift television,362 or practice
other fair uses.363
There are fewer reasons why the general public needs tools that are
“primarily useful
for the
surreptitious interception of
communications” — the tools prohibited under § 2512.364 People in
messy divorces and whistleblowers may need to surreptitiously record
audio conversations, and network systems administrators and
concerned parents may need to monitor computer communications,
but these people can use general purpose tools — tiny voice records
and network packet sniffers — that do not fall within the prohibition.
Unlike the DMCA, § 2512 seems narrowly targeted at devices like
transmitters hidden in calculators365 and specific forms of spyware.366
If regulators are bent on keeping Superusers and script kiddies
apart, perhaps they should try to model laws after § 2512 rather than
the DMCA. If one characteristic of a tool is especially pernicious and
unlikely to be useful for widespread, legitimate use, a narrow law can
be written criminalizing the creation or distribution of that tool.
CONCLUSION
Fear is with us for the long haul. With any technology as complex
and as relied upon for so many different commercially important uses
as the Internet, some will exploit disparities in knowledge to gain
power to harm others, sparking uncertainty and fear. Opportunists
will take advantage of this fear for personal and institutional gain. To
date, the fear mongers have had the upper hand, shaping policy
through sound bites and unfounded anecdotes.
Even if unchecked, the fear mongers will not spell the end of the
Internet. I am not predicting an information apocalypse. But if they
continue their stranglehold on policymaking debates, they will
eventually shape the future Internet. Policymakers will impose new
362
See generally Sony Corp. v. Universal City Studios, 464 U.S. 417 (1984)
(discussing time shifting); Recording Indus. Ass’n of Am. v. Diamond Multimedia
Sys., Inc., 180 F.3d 1072 (9th Cir. 1999) (discussing space shifting).
363
17 U.S.C. § 107 (2000); see Cohen, supra note 34, at 177.
364
18 U.S.C. § 2512 (2000).
365
See United States v. Biro, 143 F.3d 1421, 1423 (11th Cir. 1998) (affirming
convictions under § 2512 for sale of transmitters hidden in wall plugs, pens, and
calculators).
366
See Press Release, U.S. Dep’t of Justice, Creator and Four Users of LoverSpy
Spyware Program Indicted (Aug. 26, 2005), http://www.usdoj.gov/criminal/
cybercrime/perezIndict.htm (announcing indictment relating to spyware designed to
masquerade as e-greeting card).
1402
University of California, Davis
[Vol. 41:1327
regulations to constrain conduct and chill expression. Witness the
steady expansion of the CFAA. Technologists will create next
generation technologies ostensibly designed to protect, but also
designed to monitor and control. Consider how improvements in
virus-blocking firewall technology can be used by repressive
governments to search for dissident speech. Given another decade to
drive policy, the fear mongers will not destroy the Internet, but they
will change it for the worse.
Of course, fear mongers exist outside this narrow viewscreen; thus,
in some ways, theirs is not the interesting story here. The truly
troubling problem lies with the experts who have abdicated their
responsibility to discover facts and assess probabilities of risk. Experts
can reclaim some rhetorical space from the fear mongers, but to do so
they will need a rigorous, focused, interdisciplinary approach.
I worry that the abdication of expertise is a canary in a coal mine,
alerting us to a broader, evolving failure of expertise in our society. As
the definition of expertise broadens and democratizes, the effects of
fear are compounded and the biases and heuristics they introduce
become more difficult to identify and intractable to root out. In a
world without experts, or where everyone is an expert, we will look
back on today’s teeth-gnashing over how laypeople ignore expert
predictions about, for example, global warming with wistful nostalgia.
This bleak prediction is but one path we can take. By exposing the
ubiquity and persistence of online fear, and by prescribing the
medicine needed to regulate intelligently in the face of fear, this
Article points the way toward a restoration of expertise, and a move to
principled, cautious, justified policymaking and debate.
Fly UP