[glossary_exclude]The European Union, with GDPR, now has privacy regulation that is consistent and pervasive and crosses state and national boundaries. The U.S., by contrast, has a piecemeal, loose patchwork of federal and state privacy regulations that at times overlap and conflict with one another. (See Figure 1). In addition, the many different government agencies and private industry groups have created “frameworks” and “guidelines” that do not have any enforcement functions, known collectively as “best practices.” These self-regulating best practice frameworks are increasingly being used by regulators as enforcement tools.
The U.S.’ patchwork system has been expanding since at least 1986 with the passage of the Electronic Communications Privacy Act (ECPA), which bridged an existing gap between electronic transfers and computer aided transfers using the Internet. Before the ECPA, U.S. federal law had been developing since 1970 to protect electronic transfers that used traditional telephony. The result is a complicated mess of Congressional acts that contain bits and pieces of relevant law implemented piece-meal as specific actions were deemed to be violations of privacy. This explains in part why lawyers and compliance consultants solicit substantial consultation fees.
Other U.S. Acts that have electronic privacy components include:
The Federal Trade Commission Act
The Financial Services Modernization Act
The Fair Credit Reporting Act
The Controlling the Assault of Non-Solicited Pornography and Marketing Act (CAN-SPAM act of 2003)
Computer Fraud and Abuse Act
This hodgepodge system of overlapping laws arguably encourages hacking and necessitates a stronger focus on cybersecurity for organizations trying to secure sensitive and personally identifiable information (PII).
Ultimately, this meant someone needed to define elements of electronic information that should be private and protected. Although it has always been assumed that personal records such as credit card numbers and Social Security numbers must be kept private on the Internet, the federal government has been slow to define PII. In 2008, the United States Federal government defined PII in the broadest terms possible: any information about an individual maintained by an agency, including (1) any information that can be used to distinguish or trace an individual‘s identity, such as name, Social Security number, date and place of birth, mother‘s maiden name, or biometric records; and (2) any other information that is linked or linkable to an individual, such as medical, educational, financial, and employment information.
This definition focuses on privacy after the fact, meaning a person must do something in cyberspace before protection of PII becomes an issue. This definition is problematic, in part because it is reactive, but also because the EU and most of the rest of the developed world view electronic privacy as a human right. In the US electronic privacy is “action-focused.” The EU defines electronic privacy as “person-focused.”
The EU approaches PII from the “person” perspective, or what actions a person could potentially undergo that would trigger PII protection. This aligns with Article 8 of the European Convention on Human Rights: “everyone has the right to respect for his private and family life.”
GDPR Article 4 defines the person first: an identifiable natural person is one who can be identified, directly or indirectly, in particular by reference to an identifier such as a name, an identification number, location number, an online identifier or to one or more factors specific to the physical, physiological, genetic, mental, economic, cultural or social identity of that natural person.
Actions an individual takes while in the digital realm, call for a broad definition of PII. Yet from the EU’s perspective PII is a catch-all approach that might give some American business headaches as they think about selling products to EU customers. The scope of person-centered privacy is illustrated in Figure 2.
Given the EU’s approach to protecting PII, American businesses will likely have issues understanding the interplay between four specific rights GDPR expressly gives EU’s citizens. These rights are:
the Right to access personal data;
the Right to be forgotten;
the Right to erasure; and,
the Right to data portability.
American businesses have great challenges ahead as U.S. consumers are awakening to the daily incursions into their privacy. A recent USA Today poll found that Americans are now more concerned about personal privacy than either healthcare or the economy.[/glossary_exclude]
Mark Driskill currently works as a freelance writer/editor. He earned a Master’s in Archives and Records Administration (MARA) degree from San Jose State University’s School of Information, and holds a BA in History from San Jose State. Mr. Driskill’s work includes entries in the recently published International Directory of National Archives and a chapter in Teaching and Learning in Virtual Environments: Archives, Museums, and Libraries (2016). He was also one of many professional contributors to the Encyclopedia of Archival Science (2015). Additionally, Mr. Driskill co-authored the Conference Paper, “Building Trust in Government through Social Media: An InterPARES Trust Research Project.” (July 2014).