The Economics of Artificial Intelligence
Page 74
tive algorithms often assume there is a hidden truth to learn, which could be
the consumer’s gender, income, location, sexual orientation, political prefer-
ence, or willingness to pay. However, sometimes the to-be- learned “truth”
evolves and is subject to external infl uence. In that sense, the algorithm may
intend to discover the truth but end up defi ning the truth. This could be harmful, as algorithm developers may use the algorithms to serve their own
interest, and their interests—say earning profi ts, seeking political power,
or leading cultural change—could confl ict with the interest of consumers.
The danger of misleading algorithms is already seen in the controversy
about how Russia- sponsored posts got disseminated in social media during
the 2016 US presidential election. In the congressional hearings held on
October 31 and November 1, 2017, lawmakers expressed the concern that
13. https:// newsroom.fb .com/ news/ 2017/ 09/ information- operations-update/ , accessed on October 19, 2017.
14. https:// www .nytimes .com/ 2017/ 10/ 30/ technology/ facebook- google- russia .html, accessed on December 18, 2017.
15. https:// www .nytimes .com/ 2017/ 09/ 07/ us/ politics/ russia- facebook- twitter- election .html, accessed on October 19, 2017. http:// money.cnn .com/ 2017/ 09/ 28/ media/ blacktivist- russia
- facebook- twitter/ index .html, accessed on October 19, 2017.
Artifi cial Intelligence and Consumer Privacy 447
the business model of Facebook, Twitter, and Google, which depends on
advertising revenue from a large user base, may hamper their willingness to
identify or restrict misinformation from problematic users.16 Because social
media users are more likely to consume information that platform algo-
rithms push to them, they may end up consuming information that hurts
them in the future.17
The same confl ict of interest has sparked concerns in price discrimina-
tion. This argument is that if AI enables a fi rm to predict a consumer’s
willingness to pay, it could use that information to squeeze out every penny
in consumer surplus. This argument is plausible in theory, but needs to be
evaluated with at least three considerations: fi rst, if more than one fi rm
can use AI to discover the same consumer willingness to pay, competi-
tion among them will ease the concern of perfect price discrimination;
second, the economics literature has long demonstrated the ambiguous
welfare eff ect of price discrimination. As long as price discrimination is
imperfect (i.e., fi rms cannot charge every consumer’s willingness to pay),
some consumers may benefi t from the practice (via lower price) while other
consumers suff er. From a social planner’s point of view, whether to encour-
age or punish AI- enabled price discrimination depends on the weights it
assigns to diff erent parts of society. Third, in the long run, AI may reduce
the operational costs within the fi rm (e.g., via a more cost- eff ective inven-
tory management system) and foster product innovations that better fi t
consumer demand. These changes could be benefi cial to both the fi rm and
its consumers.
A somewhat opposite concern is that AI and other predictive technology
are not 100 percent accurate in their intended use. It may not introduce
much ineffi
ciency or wasteful eff ort if Netfl ix cannot precisely predict the
next movie I want to watch, but it could be much more consequential if the
US National Security Agency (NSA) fl ags me as a future terrorist based on
some AI algorithm. As Solove (2013) has argued, it is almost impossible for
someone to prove that they will not be a terrorist in the future. But at the
same time, they may be barred from air travel, have personal conversation
with friends monitored, and be restricted from work, trade, and leisure ac-
tivities. If this AI algorithm applies to a large population, it could do a lot
of harm even if the probability of error is close to zero.
To summarize, there is a real risk in privacy and data security. The magni-
16. The full video and transcript of these hearings are available at c- span .org (https:// www
.c- span .org/ video/ ?436454– 1/ facebook- google- twitter- executives- testify- russia- election- ads, and https:// www .c- span .org/ video/ ?436360– 1/ facebook- google- twitter- executives- testify
- russias- infl uence- 2016-election&live).
17. Note that a predicative algorithm is not necessarily more biased than human judgment.
For example, Hoff man, Kahn, and Li (2018) study job- testing technologies in fi fteen fi rms.
They fi nd that hires made by managers against test recommendations are worse on average.
This suggests that managers often overrule test recommendations because they are biased or mistaken.
448 Ginger Zhe Jin
tude of the risk, and its potential harm to consumers, will likely depend on
AI and other data technologies.
18.3 How Does the US Market Cope with
the Risk in Privacy and Data Security?
Before we jump into a regulatory conclusion, we must ask how the market
copes with the risk in privacy and data security. Unfortunately, the short
answer is that we do not know much. Below I describe what we know on the
demand and supply sides, along with a summary of existing public policies
in the United States. Admittedly, the literature cited below is more about
privacy and data security than about AI. This is not surprising, as AI has
just started to fi nd its way into e-commerce, social media, national security,
and the internet of things. However, given the ongoing risk and the potential
interaction of AI and that risk, it is important to keep in mind the big picture.
18.3.1 Consumer
Attitude
On the demand side, consumer attitude is heterogeneous, evolving, and
sometimes self- confl icting.
When surveyed, consumers often express serious concerns about privacy,
although self- reported value of privacy covers a wild range (see the sum-
mary in Acquisti, Taylor, and Wagman [2016]). However in real transactions,
many consumers are willing to give away personal data in exchange for a
small discount, free services, or a small incentive such as a pizza (Athey,
Catalini, and Tucker 2017). This confl ict, which some referred to as a “pri-
vacy paradox,” suggests that we have yet to comprehend the link between
consumer attitude and consumer behavior. Furthermore, researchers have
found that privacy preference varies by age (Goldfarb and Tucker 2012),
by time (Stutzman, Gross, and Acquisti 2012), and by context (Acquisti,
Brandimarte, and Loewenstein 2015). Although old data are shown to add
little value to search results (Chiou and Tucker 2014), biometric data such
as fi ngerprint, facial profi les, and genetic profi les can be much longer lasting
(Miller and Tucker, forthcoming). Hence, consumers may have a diff erent
preference on biometric data than on the data that gets obsolete fast. These
heterogeneities make it even harder to paint a complete picture of consumer
attitude and consumer behavior about privacy.
A similar puzzle exists for attitudes toward data security. A recent sur-
vey by the Pew Research Center suggests that many people are concerned
/> about the safety and security of their personal data in light of numerous
high- profi le data breaches (Pew Research Center 2016). However, accord-
ing to Ablon et al. (2016), only 11 percent stopped dealing with the aff ected
company and 77 percent were highly satisfi ed with the company’s postbreach
response.
It is hard to tell why consumers are willing to give away data in real trans-
Artifi cial Intelligence and Consumer Privacy 449
actions. One possibility is that consumers have a large or even hyperbolical
discount for the future, which motivates them to value the immediate gains
from the focal transaction more than the potential risk of data misuse in the
distant future. Other behavioral factors can be at play as well. Small incen-
tives, small navigation costs, and irrelevant but privacy- reassuring infor-
mation can all persuade people to relinquish personal data, according to a
recent fi eld experiment (Athey, Catalini, and Tucker 2017).
It is also possible that news coverage—on data breaches and privacy
problems—raises consumer concern about the overall risk, but they do not
know how to evaluate the risk specifi c to a transaction. Despite heavy news
coverage, people may have an illusion that hacking will not happen to them.
This illusion could explain why John Kelly, the former Secretary of Home-
land Security and White House chief of staff , used a compromised personal
phone for months.18
The third explanation is that consumers are fully aware of the risk, but
given the fact that their personal data has been shared with many fi rms and
has likely already been breached somewhere, they believe the extra risk of
sharing the data with one more organization is small. Survey evidence seems
to lend some support to this conjecture. According to the Pew Research
Center (2016), few are confi dent that the records of their activities main-
tained by various companies and organizations will remain private and
secure. A vast majority (91 percent) of adults agree that consumers have
lost control of how PII is collected and used by companies, though most
think personal control is important. Moreover, 86 percent of internet users
have taken steps to remove or mask their digital footprints, and many say
they would like to do more or are unaware of tools they could use.19
Consumer anxiety may explain why identity theft protection service has
become a $3 billion industry (according to IBISWorld).20 However, a market
review by the Government Accounting Offi
ce (GAO 2017) shows that iden-
tity theft services off er some benefi ts, but generally do not prevent identity
theft or address all of its variations. For instance, these services typically do
not address medical identity theft or identity theft refund fraud. In fact, a
number of identity theft service providers were caught making deceptive
marketing claims,21 casting doubt on whether such “insurance- like” services
are the best way to safeguard consumers from identity theft.
18. https:// www .wired .com/ story/ john- kelly- hacked- phone/ , accessed on October 15, 2017.
19. “The state of privacy in post- Snowden America” by the Pew Research Center, source: http:// www .pewresearch .org/ fact- tank/ 2016/ 09/ 21/ the- state- of-privacy- in-america/.
20. https:// www .ibisworld .com/ industry- trends/ specialized- market- research- reports
/ technology/ computer- services/ identity- theft- protection- services .html, accessed on October 26, 2017.
21. For example, in September 2012, Discover settled with the Consumer Financial Protection Bureau (CFPB) and the Federal Deposit Insurance Corporation (FDIC) with $200 mil-
lion refund to consumers and $14 million penalty. The CFPB and FDIC alleged that Discover engaged in misleading telemarketing on identity theft protection, credit score tracking, wallet
450 Ginger Zhe Jin
18.3.2 Supply Side Actions
Statistics from the supply side are mixed, too.
Thales (2017b) conducted a global survey of 1,100+ senior security execu-
tives, including 100+ respondents in key regional markets in the United
States, United Kingdom, Germany, Japan, Australia, Brazil, and Mexico,
and in key segments such as federal government, retail, fi nance, and health
care. It fi nds that 68 percent of survey respondents have ever experienced a
breach, while 26 percent experienced one last year. Both numbers rose from
2016 (61 percent and 22 percent).
For fi nancial services in particular, Thales (2017a) fi nds that fi rms are
aware of the cyber risk they face but tend to deploy new technology (e.g.,
cloud, big data, internet of things) before adopting security measures to
protect them. Only 27 percent of US fi nancial services organizations said
to feel “very” or “extremely” vulnerable to data threats (the global average
is 30 percent), despite the fact that 42 percent of US fi nancials had been
breached in the past (the global average is 56 percent). Consistently, both
US and global fi nancials rank data security at the bottom of their spending
plans, citing institutional inertia and complexity as the main reasons. These
numbers should be concerning because the fi nancial sector has the highest
cost of cyber crime according to the latest report from Accenture (2017).
To add a little comfort, Thales (2017a) also reports that security spending,
which includes but is not limited to data security, continues to trend up: 78
percent of US fi nancials reported higher spending than last year, trailing
only US health care (81 percent) and ahead of the overall global average
(73 percent).
Firms’ willingness to invest in data security is partially driven by the cost
they suff er directly from data breaches. A strand of literature has studied the
stock market’s response to data breach. While results diff er across studies,
the general fi nding is that the fi nancial market response is small and tempo-
rary, if negative at all (Campbell et al. 2003; Cavusoglu et al. 2004; Telang
and Wattal 2007; Ko and Dorantes 2006). A couple of studies have pro-
vided an absolute estimate of the cost. According to Ponemon (2017), who
surveyed 419 organizations in thirteen countries and regions, the average
consolidated total cost of a data breach is $3.62 million. Ponemon (2017)
further fi nds that data breaches are most expensive in the United States,
with the average per capita cost of data breach as high as $225. In contrast,
Romanosky (2016) examines a sample of 12,000 cyber events, including but
protection, and payment protection (http:// money.cnn .com/ 2012/ 09/ 24/ pf/ discover- penalty
- telemarketing/ index .html). In December 2015, LifeLock agreed to pay $100 million to settle FTC contempt charges for order violation. The 2010 court order requires the company to
secure consumers’ personal information and prohibits the company from deceptive advertising in identity theft protection services (https:// www .ftc .gov/ news- events/ press- releases/ 2015
/ 12/ lifelock- pay- 100-million- consumers- settle- ftc- charges- it- violated).
Artifi cial Intelligence and Consumer Privacy 451
not limited to data breaches. He fi nds that the cost of a typical cyber incident
(to the aff ected fi rm) is less than $200,000, roughly
0.4 percent of the fi rm’s
estimated annual revenues.
Thousands or millions, these estimates only refl ect the direct cost of the
cyber event to the fi rm, not all the consequential harm to consumers. For example, most breached fi rms off er one year of free credit monitoring service
for aff ected consumers, but data misuse can occur after a year. Either way,
consumers have to spend time, eff ort, and money to deal with identity theft,
reputation setback, fraud, blackmail, or even unemployment as a result of
a data breach. The lawsuit between the Federal Trade Commission (FTC)
and Wyndham Hotel and Resort gives a concrete example. Wyndham was
breached multiple times in 2008 and 2009, aff ecting more than 619,000
consumers. Before reaching a settlement, the FTC alleged that fraudulent
charges attributable to the Wyndham breaches exceeded $10.6 million.22
Although the fi nal settlement involves no money, this case suggests that
harm to consumers—via an increased risk of identity theft and the costs to
mediate the risk—can be much more substantial than the direct loss suff ered
by the breached fi rm. Arguably, it is this diff erence that motivates fi rms to
overcollect data or use lax data security, despite the real risk of data breach.
The good news is that market forces do push fi rms to respect consumer
demand for privacy and data security. For instance, Facebook profi les
expand over time and therefore the same default privacy setting tends to
reveal more personal information to larger audiences.23 In September 2014,
Facebook adjusted its default setting of privacy from public posting to
friend- only posting, which limits third party access to new users’ Facebook
posts. In the meantime, Facebook made it easier for existing users to update
their privacy settings, block out ads, and edit their ad profi les.24 We do not
know the exact reason behind the change, but it could be related to a few
things: for example, user willingness to share data on Facebook dropped
signifi cantly from 2005 to 2011 (Stutzman, Gross, and Acquisti 2012), aca-
demic research shows that it is very easy to identify strangers based on
photos publicly posted on Facebook (Acquisti, Gross, and Stutzman 2014),
and it costs Facebook $20 million to settle a class action lawsuit regarding
its “sponsored stories” (an advertising feature alleged to misappropriate