Privacy
Introduction
Though it brings us many benefits, the march of technology makes
an encompassing surveillance network seem almost inevitable, and
radically changes our expectations of privacy. We owe many of the
expectations of privacy we used to enjoy to a combination of immature
technology and insufficient manpower to monitor us. But these protective
inefficiencies are giving way to technologies of data processing
and digital surveillance that will change our beliefs about privacy.
We are widely tracked by public surveillance cameras, our credit-card
transactions, our passes through the fast-lanes at toll booths,
and our cell phone calls to name only a few of the data points that
could be assembled about our lives.
Each year brings more sensitive and widespread sensing devices,
including cameras, microphones, and, potentially, biological sensors,
all increasingly connected through efficient networks to ever more
powerful data processing and storage technologies. Cameras are spreading
like kudzu -- in toll plazas, on public streets, and in public parks.
We welcome them as crime-fighters, even as they eliminate our ability
to move through the world untracked. Face and voice recognition
software may soon permit image data from surveillance cameras to
be cross-referenced to profiles of each person observed. To get
a hint of the future, enter your street address at globexplorer.com.
You will see a satellite picture nearly good enough to show a car
parked in your driveway, or in mine. Better resolution is coming
soon. We are moving toward a transparent society in which our actions
and transactions are followed, our lives tracked and documented,
by folks we neither know nor trust; each of us a star in our own
Truman Show.
Two privacy writers, Simson Garfinkel and David Brin, each in their
own way, suggest the breadth and immediacy of technology's threats
to privacy. Garfinkel, the author of "Database Nation,"(1) describes a thwarted effort in the late sixties
to establish a massive central database of citizen information,
a national data center to be administered and controlled by the
United States government. Fears of Big Brother killed this proposal.
Garfinkel, who fears Big Brother as much as anyone, surprisingly
regrets the defeat of the proposal. Why? Because the alternative
that emerged in its place seems to him much more threatening. Instead
of a single known giant database in government hands, which might
have been subject to some privacy controls, we have instead many
databases in corporate hands which are not subject to privacy controls,
and which are difficult even to inventory.
David Brin, in his book "The Transparent Society,"
(2) suggests the dimension of threatened privacy
loss in another way, by the question he addresses. He poses a hypothetical
based on the assumption that ubiquitous surveillance is coming.
Imagine two cities of the future, he challenges. They look very
much the same. Each is clean, orderly, utterly without crime, with
surveillance cameras on every building and street corner. But they
are very different underneath. In one city all of the surveillance
cameras connect to the police stations where they are monitored
by government authorities. In the other city the surveillance cameras
connect to a net that allows anyone to watch. Brin asks, in which
city would you rather live? He argues that we will actually be better
off if all of us can watch each other rather than entrusting the
monitoring function to an all-powerful agency.
But must we give in to the idea of ubiquitous surveillance? It
is true that there are extraordinarily strong forces pushing in
the direction of ubiquitous surveillance, especially after the date
of 9/11/01 was burned in all of our memories. Government and business,
the two strongest forces shaping technological development of the
net, each seem aligned in their hunger for large databases containing
detailed information about us. Reexamination of our intelligence-gathering
infrastructure is well-underway, giving government access to more
and more data, and business continues to gather enormous amounts
of transactional information to aid in marketing. Is resistance
to the changes in our understandings of privacy feasible, or must
we accept a constantly examined life as inevitable?
Scott McNeely, CEO of Sun MicroSystems, asserts that we already
have "zero privacy." His advice, "Get over it."
Recent surveys, however, would seem to indicate that Americans are
not prepared simply to "get over it." Concern for loss
of privacy seems to be widespread and growing. Surveys by the Pew
Foundation (3) and Harris
Polls(4) show that Americans
want a presumption of privacy in their daily lives. Eighty-six percent
support "opt-in" requirements for Internet companies.
Fifty-four percent feel that the tracking of users by web sites
is harmful. Sixty-one percent of online-users are concerned that
their e-mail will be read by someone else, without their knowledge
or consent. Seventy-eight percent of on-line shoppers are concerned
that when buying something on-line, the personal information they
provide will result in their being sent unwanted information. Eighty-nine
percent would be uncomfortable with a web site using their browsing
and shopping practices to create a profile linked to their names.
Ninety-two percent would feel the same about a web site selling
their information to others.
Yet widespread and growing concern for privacy does not easily
translate to action – a study commissioned by the New York Times
after changes to Yahoo’s privacy policy showed that few users actively
changed their preferences, though many may have “voted
with their feet.”
Self-help
One might expect that those of us who are concerned about our privacy
would take advantage of whatever means are available to us to protect
it. A few moments of reflection about our daily routines yields
a number of privacy-enhancing adjustments that we could make. When
using the Internet, for example, we could employ different identities.
We could disable cookies, use multiple e-mail addresses, and fake
the personal data sought by web sites. We could stay off the Internet
all together. When in public, we could take steps to avoid being
"captured" by surveillance cameras-refrain from using
ATMs, avoid those businesses and public places where cameras are
installed, or just stay home. We could take advantage of direct
marketing opt-out lists by sending a letter or making a phone call.
We could avoid credit cards in favor of using cash, telephone from
pay phones, and avoid automatic toll lanes. We could encrypt our
electronic messages and our files, and route our communications
through anonymizers. With enough planning and effort, we might avoid
the surveillance net.
Yet despite widespread and growing concern for privacy, most people
are unable or unwilling to take advantage of privacy-enhancing strategies
that are open to us. Like most visitors to our national parks who
idealize the beauty and isolation of the great outdoors without
ever straying from the well-trodden trails and asphalt roads, most
of us express concern for our loss of privacy but do little to protect
it.
That there are costs for protecting privacy does not of course,
in and of itself, preclude the use of privacy enhancing techniques. We do, after all, incur costs
all the time for things we desire. We buy and install blinds for
our bedroom windows. We build fences in our yards. The problem is
that when it comes to privacy from electronic intrusion, the costs
of protection are often higher than the perceived benefits.
Professor Michael Froomkin explains this in terms of what he calls "privacy
myopia."(5) Individuals tend to undervalue the bits of
information about themselves that allow aggregators to build databases
and profiles. The problem, Froomkin says, is that any one bit of
information about ourselves doesn't seem that valuable to us. This
being so, it doesn't seem worth it to go to the trouble of protecting
it. Yet, to the aggregator, to the person who is on the other end
putting together the profile that will be sold and used as an asset
in marketing, that bit of information has value. From the individual's
point of view, it is hard to see what the aggregator sees, and hard
to know the uses to which the information will be put. This leaves
the individual willing to give the information away or sell it cheap.
The effect of this privacy myopia is that in a given transaction,
too often the costs of withholding, in the interest of privacy,
will appear higher than both the values of the information withheld
and of the privacy gained from withholding it. Such seemingly rational
cost avoidance is even more apparent when deciding whether or not
to purchase privacy enhancing products. As one understandably disillusioned
cryptographer (Bruce Scheiner) put it: "Privacy never seems
to sell. Those who are interested in privacy don't want to pay for
it."
Judicial Protection
Courts have traditionally protected privacy primarily in its connection
with real property, enforcing the law of trespass to give us privacy
behind fences and closed doors, protecting privacy of place but
not privacy of presence. The Supreme Court made a great theoretical
advance in privacy law when it announced in Katz
v. United States(6) that
the Constitution protects people, not places. While the law is that
our reasonable expectations of privacy, in addition to the physical
bounds of our real property, are to be protected, the traditionalist
roots of privacy doctrines still see the home as the paradigmatic
place that privacy is protected, approaching public surveillance
as a risk commonly assumed by people operating in the modern world.
Grounding privacy in reasonable expectations might seem to offer
great promise for privacy protection in the face of advancing technology
because of the adaptability inherent in the standard.
But expectation can be an insecure footing for privacy, like
sand slipping out from under foot as one climbs a dune. The Supreme
Court’s 2001 decision in United
States v. Kyllo,(7) provides
both encouragement that a more nuanced approach could be taking
hold and caution that traditional notions of property-based privacy
rights govern. In Kyllo, authorities used a thermal imager to look
through the walls of a house. Thermal imagers collect molecular
information in the form of heat that radiates through walls. Justice
Scalia, during the oral argument asked the petitioner a cogent question:
"Why don't your reasonable expectations of privacy include
technology?" Inasmuch as there are thermal imagers in the world,
why not expect people to guard against them just as "you pull
your curtains if you want privacy because you know people have binoculars?"
Then, to the suprise of many, Justice Scalia wrote the majority
opinion for a five-to-four court holding the thermal image scan
to be an unlawful search. He emphasized the expectation of privacy
in the home from sense-enhancing technologies, at least where the
technology is not "in general public use." What this means
for online privacy remains to be seen. Connections through the Internet
may originate from one's home, but are not confined to it. People
may expect privacy in their online activities, but that has been
misconception rather than reality.
Where new technology allows capture of information that was previously
not subject to capture, expectations of privacy are likely to be
considered naïve and uninformed, grounded in immature technology
rather than in law. Claims of privacy intrusion resulting from aerial
surveillance of one's house, or thermal imaging of heat emanating
from one's walls, or radio pickup of cordless telephone conversations
have all failed because the expectation of privacy was not deemed
to be objectively reasonable. Effectively, the courts seem to say
that, as far as the Constitution is concerned, people's expectations
of privacy must change to adjust to the capabilities of new technologies.
Legislative Protection
The ultimate legislative goal of many privacy advocates in the
United States is to have privacy seen as a human right of overarching
importance, and not merely a customer relations problem. They point
to the European Union and its long-established Data
Protection Directive as the example of the right way to approach
the problem. The basic elements of such privacy protection
would include requirements of notice, informed consent, access to
profiles, availability of process to correct errors, limitation
of the secondary uses to which gathered information may be put,
and a well-funded agency charged with promoting the values of privacy
and enforcing the laws that protect it, something approaching an
environmental protection agency for privacy.
So far, industry has successfully resisted legislative movement
towards a human-rights approach to privacy in the U.S., though many
states are beginning to act to impose requirements in areas where
the federal government has feared to tread. Minnesota has enacted a broad law governing the collection and use
of data by Internet service providers, and the California State
Assembly recently passed a bill requiring privacy policies and basic
standards for outlining online practices to consumers. Industry
objects to such privacy regulation as antagonistic to the First
Amendment, almost un-American and offers self-regulation as the
alternative.(8) The Wall Street Journal reports that
new privacy regulation will cost businesses as much as $36 billion.(9)
Privacy advocates see the problems with self-regulation
as fundamental: industry's objective in self-regulation is not primarily
to protect privacy, but rather to allay customer fears sufficiently
to do business, and to hold government regulation at bay. In the
long run, industry self-regulation is a shifting ground for privacy
protection, just as is the expectation of privacy to which courts
respond.
Notes
(1) Simson Garfinkel, Database
Nation, The Death of Privacy in the 21st Century, O'Reilly (2000).
(back to text)
(2) David Brin, The Transparent Society, Addison-Wesley
(1998). (back to text)
(3) Pew Internet & American Life Project, Trust and Privacy
Online: Why Americans Want to Rewrite the Rules (2000) (back to text)
(4) BusinessWeek/Harris Poll, A Growing Threat
(March 2000) (back to text)
(5) A. Michael Froomkin, The
Death of Privacy?, 52 Stan. L. Rev 1461, 1471 (2000) (back to text)
(6) Katz v. United States, 389 U.S. 347 (1967)
(back to text)
(7) Kyllo v. United States, 2001 U.S. Lexis
4487 (June 11, 2001). (back to text)
(8) Online
Privacy Alliance (back to text)
(9) Internet
Privacy Rules Could Cost Businesses as Much as $36 Billion,
The Wall Street Journal Tech Center (May 2001)
Back
to Top | Intro | Case Studies | Discussion Topics | Resources
Case Study 1: Putting Court Records Online
While many have rightly focused on threats to privacy posed by
the constant data collection that goes on through Internet Protocol-based
online networks, there are massive databases that contain incredible
secrets about many Americans – the public records filings contained
in court papers. Criminal
records, allegations in nasty divorce suits, accusations of negligence
and intentional neglect, and massive amounts of personal information
are present in court records across the nation.
With the drive to introduce e-government and make interacting with
the various governmental bodies across the nation easier, many jurisdictions
are looking to put their court records online, searchable from anywhere.
Appreciating the potential problems that this could cause,
numerous lawyers, judges, and privacy activists are resisting the
move.
There are clearly situations where harm could come from the increased
availability of information – identity theft facilitated by the
harvesting of names and social security numbers, abused spouses
tracked down through a listing of their address on an application
for a protective order, property transfer and other financial fraud
made easier through the massive collation of records. How do we fulfill the promise of open records and transparent government
while managing the risks that such information could be misused?
Case Study 2: Spyware - How Private Is
the Web?
Recent revenue pressures on both traditional e-commerce companies
and peer-to-peer shared software has created a new dimension to
web surfing. More and more often, when individuals download software
they receive more than they expected. Unbeknownst to them, “Spyware”
or adware has been bundled with the product they download. Such
software allows companies to trace the user's online movements and
potentially report the information back - all without the user's
knowledge or consent.
Users rarely realize Spyware programs are bundled with what they're
getting because the programs are automatically loaded during installation.
Although some companies do alert users to what they are installing,
either in the licensing agreements or in the privacy section, few
users actually read the notice, and if they do, the information
provided may not fully explain what the spyware application will
do.
Spyware works by embedding a unique identifier onto your computer.
It remains even after the application it came with is removed and
constantly runs in the background on your computer. It enables companies
to track all of the user's online movements, and may report that
information back to a central server.
Spyware has been identified primarily with file sharing services
like Kazaa, Gnutella BearShare and Audio Galaxy Satellite, but also
is present in RealNetworks software (attached to RealJukeBox), many
browser extensions, and CometCursor. It has even been found attached
to cyber-nanny programs like SurfMonkey and on the Learning Company
CD-ROMs.
These programs have understandably created a controversy but have
been justified by some as the price for freeware. In a CNET article
Vinnie Falco, the CTO of FreePeers, candidly recognized that "One
of the issues around free software is the need to make money somehow.
[Spyware] is a great compromise between protecting user privacy
and the ability to support free software." The efforts of some
companies have been very successful in this area. For example, one
ad company, Radiate , (as of
6/7/02, site down) has succeeded in placing spyware on 30 million
PCs. That is a huge market base for any retailer wanting an easy
marketing solution, but can it be justified?
As one author lamented, the basic desire to increase revenues through
targeted advertising can become more problematic. "Theoretically
speaking, advertising-supported software's capabilities are not
just limited to collecting demographic data or reading your browser
history. They can explore your computer, randomly scan files and
ferret out sensitive information like your credit card number."
(Subha Vivek, "Are You Being Watched." Frost.com, 5/18/01)
Even if companies don't go that far, the software still affects
the performance of your computer.
Recent focus on the issue has led some companies to change their
practices. Although The Learning Company claimed that it does not
gather personal information from its software, it added a program
to former parent company Mattel’s website that will uninstall the
unwanted programs from people's hard disks. Similarly SurfMonkey
and RealNetworks indicated that they too would stop sending personal
information back to its servers. However, given the obvious success
and usefulness of spyware, it is not likely to disappear.
If you're interested in finding out if your computer is running
spyware, use the resources below:
- SpinRite offers OptOut, a free software
that cleans your system registry of spyware;
- Ad Aware from LavaSoft,
a free utility that scans your memory, registry and hard drive
for spyware;
- SpyWare lookup database allows you
to check whether the free software you are about to download
comes with spyware.
Back to Top | Intro | Case Studies | Discussion Topics | Resources
Case Study 3: Information and Threats
One of the most contentious issues in American politics is the
scope of reproductive rights. It should come as a surprise to no one that
information technologies, including the Web and increasingly inexpensive
and easy to use video technologies, are used by people of varying
perspectives to advance their cause.
Whether you characterize their views as “pro-life” or “anti-choice,”
those who look to restrict access to abortion worldwide are using
information technologies to amplify their message.
The Nuremberg Files website www.christiangallery.com/atrocity
aims to gather information about those who perform and seek abortions Those associated with the site have been linked
to many aspects of anti-clinic violence, including the murders of
three doctors who performed abortions, and the site has advocated
further crimes.
In the recent 6-5 en banc decision in Planned Parenthood v. American
Coalition of Life Activists, the Ninth Circuit Court of Appeals
upheld a jury verdict that certain “Wanted” posters about doctors
who perform abortions and associated personal information in context
constituted a “threat of force” under the Freedom of Access to Clinic
Entrances Act.
As the Wall Street Journal recently reported, the new tactic of
clinic protestors, including the Nuremberg Files site, is to take
pictures of women going into reproductive health clinics and posting
them on the Web. On the
Nuremberg Files site, the project is described as follows: “People will locate themselves outside baby butcher businesses
across the nation and film the people coming and going. We presently
have a group of sidewalk counselors who are committed to launching
the project. We want to connect with every baby butcher business
in the nation so we can catalog the people who go out to kill God's
little babies.” Where possible, activists seek to
obtain as much personal information about the woman in question
and also put it up online.
At what stage does the posting and dissemination of personal information
constitute a threat? What
do you see as the implications for society in allowing such information
dissemination in a heated context – facilitating speech, but also
potentially facilitating awful crimes?
What are the implications of restricting it?
Back to Top | Intro | Case Studies | Discussion Topics | Resources
Discussion Topics
1. Do you care about
your privacy online? What do we have to worry about? Is the concern
worth the trouble of worrying about it? Of doing something about
it?
2. Are there differences
between worries about privacy online and that in the offline world? Where the two merge, as in the Nuremberg Files
case, what should we do?
3. How do you
explain the apparent anomaly between the high measure of expressed
public concern for online privacy and the low willingness of people
to use self-help tools to protect it?
4. What, if any, protection
should we expect from courts for our privacy online? Are courts
capable of providing it?
5. Would you favor the
establishment of a privacy commission?
6. Enhanced 911
and presence detection technology (otherwise known as instant messaging)
will allow for greater ability to pinpoint cell phone users at any given time. This
obviously has great benefits in assuring that emergency services
can reach people on cell phones, and in finding friends on a network. But there are obvious downsides as well – you
may not always want to be found, and the business imperative to
use the technologies for marketing purposes could turn your phone
into the latest repository for spam.
How would we attempt to harness the benefits of the technology
while mitigating potential harm through a legally-based strategy?
A technologically-based one?
An economically-based one?
Back to Top | Intro | Case Studies | Discussion Topics | Resources
Privacy Toolbox
Cryptography
A basis of online privacy tools is the science of cryptography.
Cryptography
is defined by TechWeb Encyclopedia as "the conversion of data
into a secret code for transmission over a public network. The original
text or 'plaintext' is converted into a coded equivalent called
'ciphertext'
via an encryption element. The ciphertext is decoded (decrypted)
at the receiving end and turned back into 'plaintext.'
" Privacy tools rely on the use of encryption devices for securing
data and data transactions, sending email, and browsing.
As adapted from The Atlantic Monthly, March 2001 "Open Secrets"
The origins of the computer came from the need, during World War
II, to create something that would be able to crack enemy code.
The scrambling of code to insure that only the intended recipient
receives the message is called cryptography. After the war, governments
-- especially the United States -- took over cryptography and made
it an absolute secret claiming to be protecting its citizens from
terrorists, hackers and other criminals. In the 1990's, however,
some "technological crusaders" who were troubled by the
government control of private information and its implications,
managed to give access to high-powered cryptography to the general
public. Since then, cryptography has existed in our every day lives
as seen through the use of ATMs, electronic banking, cell phones
and web browsers. The impact of cryptography is tremendous according
to Lawrence Lessig and others who believe that without it people
will not be able to protect their private, personal information
online.
Currently the invasion and protection of privacy is an important
issue. Cryptography, although will create new crimes, it will stop
others. (according to Lessig's "Code" p. 36). "[Cryptography]
prevents eavesdroppers from listening to our conversations and reading
our e-mails, and from reading information stored on hard disks connected
to the Internet." Privacy-protection companies such as Zero-knowledge
Systems and Lumeria base their business models on cryptography.
In order to use these tools effectively, however, there must be
an element of trust present between the customer and the cryptosystem.
This is the only way to have faith that your private data will be
well protected.
Cookies
According to TechEncyclopedia, a cookie
is defined as "Data created by a Web server that is stored
on a user's computer. It provides a way for the Web site to keep
track of a user's patterns and preferences and, with the cooperation
of the Web browser, to store them on the user's own hard disk."
The purpose of a cookie is twofold, with benefits to both consumers
and sellers: It enables a consumer or user to have her information
remembered by a site. For sellers and advertisers, cookies allow
tracking of consumers' habits to target specific advertisements
to users with certain buying and surfing habits. For a review of
data collection technology, see Rita Lin, Information
Collection, in E-Commerce: An Introduction (2001) and Internet Cookies, Information
Bulletin I-034, U.S. Department of Energy Computer Incident Advisory
Center (1998).
Try out Brian Wells'
Cookie Demonstration
For more information on cookies, see Electronic Privacy Information
Center, The Cookies Page.
For a great demo of cookies, see Privacy.net, Bake your own Internet Cookie.
To find out who's tracking you with web bugs (pixel-sized cookies),
use the Privacy Foundation's Bugnosis.
For a demo of how advertising companies like DoubleClick operate
cookies with banners, see Privacy.net, How
Companies Can Track Your Movements on the Internet.
Anonymizer
Anonymizer is an anonymous browsing service.
This technology blocks cookies, Java, and Javascript; encrypts cookies,
email, Web addresses (URLs) in the user's browser history list but
not the page titles; and allows the user to chat and browse the
Internet while concealing the user's identity from their ISP. This
tool rewrites the links on each Web page the user visits, thus requiring
no browser reconfiguration. With its 'Safe Cookies'
encryption device, the Anonymizer enables cookies from first party
sites that require them. It allows third party cookies for sites
where the user wants to use them and then immediately disables the
cookie and eliminates it from the user's browser after the surfing
session.
Microsoft Internet Explorer 6 & P3P Technology
Microsoft Internet
Explorer 6 (IE6) includes a cookie filtering device imbedded
in the browser. The device allows the user to choose from among
six privacy preference settings (from the "Block All Cookies"
setting to "Accept All Cookies" setting). The user may
also opt to import privacy settings from other sites. IE6 previews
a site's privacy policy in regard to cookies, provides the user
with a one-line description of the policy (the "compact policy"),
and informs the user of whether a site collects personal identifiable
information and whether a site's compact policy states if this information
will be used for secondary purposes. The browser also scans the
policy for an opt-in or opt-out provision. The browser reads only
policies that are written in P3P (Platform for Privacy Preferences Project)
XML code, a format developed by the World Wide Web Consortium.
Zero-Knowledge
Zero-Knowledge Freedom is an anonymizing
tool that provides firewall protection, filters cookies, automatically
fills out online forms at user's discretion, manages and blocks
online ads, monitors outgoing text and alerts user before sending
personal identifying/sensitive information, encrypts email, and
allows the user to browse and chat on the Internet anonymously.
Freedom 2.0 allows the user to choose from among multiple (up to
five) user-created actual or pseudonymous digital identities (called
"nyms") to use for filling out forms, browsing, chatting,
and sending email. A note on firewalls: A firewall
is "a method for keeping a network secure." The Privacy
Foundation glossary decribes it as "a software or hardware
device to control access to computer on a local area network (LAN)
from outside computers on the Internet."
Opt-In/Opt-Out
Opt-in and opt-out are two options that a user may be provided
when deciding whether to receive, send, or accept information, communications,
or a service. Some web sites include an opt-in or opt-out provision
in their privacy policies. To opt-in
is "(t)o purposely accept some situation or condition ahead
of time." To opt-out
is "(t)o cancel some situation or condition." For example,
under an opt-in provision, a web site may share a user's information
with other businesses only if the user requests or gives
the web site permission to do so. In the same situation under an
opt-out provision, the web site may share the user's information
with other businesses unless and until the user requests that the
web site discontinue doing so. These provisions may be activated
by sending an email to the web site, or by clicking or checking
a box on a page of the web site.
Back to Top | Intro | Case Studies | Discussion Topics | Resources
Additional Resources
FindLaw Online
Privacy Law
Washington Post, Scott McNealy, “The Case
Against Absolute Privacy,” May 29, 2001. Discusses consumer
benefits of sharing information online. Advocates self-regulatory
approach of businesses to online privacy.
“The Emergence of Website
Privacy Norms,” Steven A. Hetcher, 7 Mich. Telecomm. Tech. L.
Rev. 97 (2000 / 2001). Analyzes history and current trends in website/consumer
relationships in exchange and use of personal data.
“The Privatization
of Big Brother: Protecting Personal Sensitive Information From Commercial
Interests in the 21st Century,” Mike Hatch, 27 Wm.
Mitchell L. Rev. 1457 (2001).Argues for an opt-in policy view in
all legislation concerning the sharing of personal data with secondary
parties.
“E-commerce Privacy and the Black Hole of Cyberspace,” Stephen R.
Bergenson, 27 Wm. Mitchell L. Rev. 1527 (2001).“The Privacy Paradox,”
Eric Jorstad, 27 Wm. Mitchell L. Rev. 1503 (2001).
“The Criminalization
of True Anonymity In Cyberspace,” George F. du Pont, 7 Mich.
Telecomm. Tech. L. Rev. 97 (2000 / 2001).
“Internet Privacy: Does
the Use of “Cookies” Give Rise to a Private Cause of Action for
Invasion of Privacy in Minnesota?,” Gregg M. Fishbein and Susan
E. Ellingstad, 27 Wm. Mitchell L. Rev. 1609 (2001).
“Window Peeping in the
Workplace:A Look Into Employee Privacy in a Technological Era,”
Donald H. Nichols, 27 Wm. Mitchell L. Rev. 1587, (2001).
“Fair Information Practices
and the Architecture of Privacy: (What Larry Doesn’t Get),” Marc
Rotenburg, 2001 Stan. Tech. L. Rev. 1 (2001). Critiques Larry Lessig’s
book Code and Other Laws of Cyberspace and its approach to
privacy protection.Working
paper version available.
International/European
Union Perspective“Internet Regulation—Heavy Handed or Light Touch
Approach?A View From a European Union Perspective,” Robert T J Bond,
27 Wm. Mitchell L. Rev. 1557 (2001).
Jessica Litman, Information Privacy/Information Property,
52 Stan. L. Rev. 1283 (2000).
Joel R. Reidenberg,
Restoring Americans' Privacy in E-commerce, 14 Berkeley Tech. LJ
771 (1999).
Lessig, Code, Chapter
11, "Privacy"
Zittrain, "What
the Publisher Can Teach the Patient", Stanford Law Review,
Vol. 52
Randy Barret, "Stealth
Chatter." ZD Ney 5/10/01
Communications Decency
Act: 47 USCA § 230
Electronic Communications
Privacy Act: 18 USCA § 2702P3P and Privacy on the Web FAQ - a project
aimed at creating a technological solution to privacy problems.
Maybe we will never need to read the privacy policy again.
EPIC, "Pretty Poor
Privacy: An Assessment of P3P and Internet Privacy" David Chaum,
"Achieving Electronic Privacy," Science (August 1992):
96-101. Scientific American, "Computer Security and the Internet,"
October 1998 (hard copy only)
The following links
offer some interesting perspectives and a few examples of experiments
gone awry.
Privacy Foundation,
"Work
Place Surveillance is the Top Privacy Story for 2000,"
(12/28/00); Statistics on work place monitoring and links to related
articles.
ECompany, You're Being
Followed (This Is Not News). December 1, 2000
Solveig Singleton, Privacy
as Censorship: A Skeptical View of Proposals to Regulate Privacy
in the Private Sector, Cato Policy Analysis No. 295, January 22,
1998.
US News & World
Reports: "More
Web Users Wage Guerilla War on Nosy Sites" (8/28/00)
FTC, "Online Profiling:
A Report to Congress," June, 2000
Industry Standard, "Toysmart
Settles with FTC," July 21, 2000
CNET, "Thirty-nine
States Object to Sale of Toysmart's Customer List," July 21,
2000
Jean Camp, Trust and
Risk in Internet Commerce. Chapter 3
Brian Sullivan, "Sen.
Kerry: On-Line Taxes and Privacy Changes Coming," Computer
World (5/18/01)
Keeping "private
e-mail" private: A proposal to modify the Electronic Communications
Privacy Act. Robert S. Steere, 33 Val. U.L.Rev. 231 (1998).
Carnivore Eats Your
Privacy, Wired News, 7/11/00
Critics Blast FBI's
First Release of Carnivore, CNET, 10/2/00
EPIC's Carnivore Archive,
Electronic Privacy Information CenterNYTimes, "Judge Sets F.B.I.
E-Mail Scanning Disclosure," August 3, 2000 (follow trail of
article links on the right).
Newsbytes.com, "Administration
Bias Alleged in Carnivore Review Team," October 4, 2000
Privacy Foundation,
"E-mail
Wiretapping," 2/5/01
CNET, "Beware:
E-signatures Can Easily Be Forged," July 14, 2000
Reuters, "Clinton
Relaxes Crypto Export Rules," July 17, 2000
1999 FTC Report
Testimony
of Joel R. Reidenberg: Professor of Law and Director of the
Graduate Program Fordham University School of Law before the House
Subcommittee on Commerce, Trade and Consumer Protection, Committee
on Energy and Commerce: Hearing on the EU Data Protection Directive:
Implications for the U.S. Privacy Debate (March 8, 2001 )
Justin Boyan, "The Anonymizer: Protecting
User Privacy on the Web," CMC Magazine (December 1997).
Toby Lester, "The Reinvention
of Privacy," Atlantic Monthly (March 2001) (includes overview
of Zero-Knowledge and Lumeria).
Massachusetts Institute
of Technology, Ethics and Law on the Electronic Frontier, "Group
Project: Protecting privacy through anonymity tools," (Spring
2001) (listing resources on privacy tools and online anonymity).
World Wide Web Consortium,
Platform for Privacy Preferences (P3P) Project..
"As Congress Mulls
New Web Privacy Laws, Microsoft Pushes System Tied to Its Browser,"
Glenn R. Simpson, Wall Street Journal, March 21, 2001.
"A Pseudonymous Communications
for the Internet," Ian Avrum Goldberg, Dissertation for
PhD in Computer Science at University of California at Berkeley,
Fall 2000. Dr. Goldberg is Chief Scientist for Zero-Knowledge Systems.
"Industry Wants to Opt
Out of Opt-In," Yair Galil, Internet Law Journal, April
16, 2001. Critiques opt-in provisions and provides several links
to studies, cases and other articles.
"Network Advertising
Initiative: Principles not Privacy," Electronic Privacy
Information Center, July 2000. Critiques opt-out provisions
Back to Top | Intro | Case Studies | Discussion Topics | Resources
|