How To Regulate Tech: A Technology Policy Framework for Online Services – Center For American Progress

Center for American Progress

See all

Center for American Progress
Online services have become an essential and ubiquitous part of American life. This report proposes a new regulatory framework to address existing harms, promote equitable growth, and protect the public interest online.
Restoring Social Trust and Strengthening Democracy, Biden Administration, Consumer Protections, Democracy, Infrastructure, Innovation, National Security, Nondiscrimination Protections, Politics and Elections, Technology Policy
Associate Director, Media Relations
Director, Media Relations
Senior Media Manager
Associate Director, Media Relations
Director, Federal Affairs
Deputy Director, State and Local Government Affairs
Vice President, Government Affairs
Director, Federal Affairs
Manager, State and Local Government Affairs
For an overview of the Center for American Progress’ full report, “How To Regulate Tech: A Technology Policy Framework for Online Services,” see the executive summary.
Read the summary
Online services have evolved from novel communication tools to ubiquitous infrastructure for the United States’ economy, democracy, and society. In 1995, 42 percent of U.S. adults said they had never heard of the internet.1 Today, 85 percent of American adults are online every day, with nearly a third online “almost constantly.”2 This evolution has driven significant economic and social growth: Internet businesses now create 10 percent of U.S. gross domestic product (GDP),3 on par with manufacturing, and nearly 15 percent of U.S. retail is e-commerce.4 A majority of U.S. adults use social media, with large proportions using social media every day,5 and many get their news through digital channels.6 During the COVID-19 pandemic, nearly 93 percent of households with school-age children used some form of online “distance learning,”7 and 71 percent of U.S. adults worked from home all or most of the time.8 Online services have created tremendous benefits and new opportunities for expression. They have become an essential part of American lives and livelihoods.
Simultaneously, the growth of online services has created new inequalities, acute consumer protection issues, and troubling concentrations of power. Online service companies have produced substantial wealth, but these gains have failed to reach the American workforce more broadly.9 Pervasive, ubiquitous digital surveillance has eroded Americans’ civil liberties. Exploitation of people’s data has created novel consumer threats around privacy,10 manipulation of consumer behavior,11 and discrimination.12 Americans face these and other harms from online services, including but not limited to widespread fraud,13 abuse of small businesses,14 abuse of market power,15 faulty algorithms,16 racist and sexist technological development,17 cybersecurity challenges,18 threats to workers’ rights,19 curtailed innovation,20 and challenges with online radicalization and misinformation.21 These harms have affected all Americans but have had a disproportionate impact on low-income people, people of color, disabled people, and other systematically targeted communities. A persistent lack of transparency has compounded the public’s ability to fully understand—let alone address—many of these challenges.
These problems are not inevitable. The economic, civil rights, privacy, and consumer protection harms from online services are not the necessary “cost” of flourishing online culture, commerce, and innovation. Rather, these harms are the result of business decisions and market failures, regulatory gaps and enforcement oversights, which have together produced an environment in which harmful and predatory practices among online services are industry standards.
Support for government action to regulate online services has grown as people’s lives and livelihoods become more dependent on the internet.22 Yet Congress has never passed a comprehensive framework for regulating online services, leaving federal oversight fragmented, incomplete, under-resourced, and unable to respond to emerging or even established harms in a timely manner. Considering the criticality of online services to not only the national economy but also the fabric of democracy and society, this level of risk has grown to be intolerable. In the robust progressive tradition of regulating industries in the public interest, wherein antitrust and regulation work together, online services too require new laws and regulatory authority backed by substantial resources. Multiple overlapping tools and specialist oversight are needed to identify and mitigate significant risks and prevent systemic problems.
Building on decades of incisive research on digital harms, experts have put forth stirring analyses of current problems and novel proposals for digital regulatory reform.23 Others have powerfully explained the need for robust antitrust action, the case for structural separation as a key tool, and the challenges ahead.24 Governments have laid the groundwork with efforts focused on pressing issues in digital marketplaces, from the U.S. House Antitrust Subcommittee’s digital markets investigation and subsequently proposed legislation to new regulatory proposals from the European Union and the United Kingdom.25
The United States now has the opportunity to reestablish a public interest vision for online services. The 117th Congress can address several immediate issues by passing the bipartisan tech antitrust package26 and taking up federal privacy legislation. It can make needed investments in agencies such as the U.S. Federal Trade Commission (FTC), which has existing authorities it can bring to bear to protect Americans online.27 Reinvigorated antitrust enforcement can also deliver significant advances in competition and privacy. These are essential steps toward restoring competition, establishing privacy rights, and correcting some of the immediate online services harms. Yet even if these measures are implemented, significant gaps will remain in the government’s fundamental ability to anticipate, understand, and address the harms from all online services—not just those with monopoly power—while balancing the multiple, competing interests at the heart of many sociotechnical regulatory decisions.
The Center for American Progress seeks to advance a dynamic conversation about what is next in technology policy regulation. This report presents a new, commonsense framework for regulating online services of all sizes. These proposals are conceived as the next phase of restoring public interest oversight online—future action that is complementary to the steps the United States should take today. They aim to address current harms, create the capacity to prevent future issues, and promote innovation in the public interest for years to come. Building on existing work, this report makes five primary contributions:
Historically, after-the-fact litigation has been too slow-moving to alone address online services harms. As Americans increasingly grapple with these harms and threats to the public interest, an ad hoc approach to online services is increasingly insufficient. To anticipate technology’s evolution and balance difficult trade-offs, regulators should have proactive rule-making abilities to curb problems before or as they occur. New statutory prohibitions of problematic online services practices are likewise required to set clear rules of the road, especially for stable, long-standing online services markets. In combination, a hybrid regulatory approach backed by substantial resources and oversight powers is needed to tackle the range of public interest issues raised by online services.
This report proposes a high-level framework for thinking about the universe for regulatory action, the goals for regulation, potential tools to accomplish those goals, and ways to structure them. It envisions many potential pathways to actualizing this vision—a combination of new and existing statutes, new rule-making powers, and revived use of existing powers. The report also emphasizes that, given the inherent administrative challenges of regulatory oversight, use of structural separation regimes and clear statutory rules are preferable where appropriate and possible. Current FTC Chair Lina Khan and other experts have argued persuasively that structural separation approaches have particular relevance to digital platform competition, especially where platforms play a dual role of marketplace operator and competitor within the marketplace.28 But this is not always feasible, especially in addressing issues that extend beyond the competitive harms from the largest players. This report conceptualizes a regulatory universe for all online services, encompassing potential structural separations, statutory prohibitions, new rule-making capacities, and increased capacity for oversight.29
This framework envisions several potential strategies for regulatory administration. Instead of preemptively determining which federal bodies should administer this approach, the authors hold that form should follow function. The report closes by discussing a set of administrative options but remains agnostic among those choices. Regardless of which agency path is chosen, significant future work will be required to overcome a range of administrability challenges that have limited the effectiveness of past efforts. Among them is designing robust safeguards against industry capture—wherein policymakers become unduly influenced by the industries they oversee—while also ensuring requisite specialized expertise. A range of technical and sociotechnical expertise—a term used to describe the blend of technical and social sciences skills needed to understand how people and technologies interact—is needed to holistically understand, anticipate, and remedy harms from online services for all Americans.

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.
In defining the universe for regulatory action, this proposal focuses simply on providers of “online services,” meaning products and services delivered through the internet. This straightforward approach acknowledges the increasingly digital nature of economic activity: Not all businesses that provide online services, for instance, are necessarily “tech companies.” Therefore, a cross-cutting approach that focuses on the online service components delivered by many different types of providers is appropriate and complementary to existing, sector-specific regulations. This universe includes cloud infrastructure, artificial intelligence (AI) services, Internet of Things (IoT) devices, algorithmic decision-making systems, online advertising, app stores, media-sharing services, operating systems, search engines, e-commerce platforms, data analytics services, social media services, and more. A focus on online services excludes core internet networking protocols—such as those managed through the Internet Engineering Task Force—as well as simple, static use of those protocols to display content, which does not rise to the level of service provision. Undoubtedly, there is a productive debate to be had on the optimal definitional approach.
This report proposes three complementary regulatory tiers to account for the wide variation in online services: online infrastructure, general online services, and gatekeepers. All online services, regardless of size, would fall into one of two categories: general online services, by default, or online infrastructure, which is opt-in and subject to regulator approval. Services that do not opt in to the online infrastructure tier are subject to any general online services rules; large, extremely powerful entities may additionally qualify as gatekeepers and face more tailored rules.
The online infrastructure tierdesigned for infrastructural products such as web hosts, cloud services, and content delivery networks—is an opt-in regulatory category that aims to preserve online infrastructure by imposing public interest obligations such as common carriage, a requirement to deal fairly and equitably with all legal customers; nondiscrimination; and cybersecurity and other standards alongside greater regulatory stability and dedicated intermediary liability protections separate from Section 230 in the event that the law is changed. It provides baseline freedom of expression protections for legal content online and would enable a more focused discussion on carefully calibrated intermediary liability changes to higher-stack, consumer-facing services.
The general online services tier is designed for all other online services entities, regardless of size. It proposes prioritizing competition, civil rights, consumer protection, and privacy as the key principles for online services regulation—operationalized by dedicated statutes and accompanying rule-making capabilities guided by those principles and any process requirements enumerated by Congress. Clear, per se violations of rules can set a foundation for online services conduct. Additional principles-based rule-making would enable regulators to sustainably update and tailor protections to keep pace with emerging markets, balance competing interests within rule-making, and curb predatory practices on an ongoing basis. Proposed process requirements incorporate consideration of information diversity and pluralism, innovation, equitable growth, and representation of all participants in multisided markets. Equipped with significant technical expertise and research capacity, these regulators would be tasked with significant general oversight responsibilities over online services and also serve as specialist partners and referrers to other federal agencies.
The gatekeeper tier provides additional oversight for the largest online services companies whose dominance and power extend beyond a specific market, similar to recent work from the EU’s Digital Markets Act and the tech antitrust package introduced in the 117th Congress. In order to determine gatekeeper status, this report builds on existing work to propose a new test of qualifying characteristics of dominant digital services. It proposes an array of specific thresholds for discussion, such as measuring significant market power through either a 30 percent or higher market share or persistently high Q ratios. For companies that qualify, this report envisions additional powers to complement existing antitrust enforcement. These include proactive rule-making powers and the ability to administer tailored remedies and sanctions, also guided by a narrow set of principles set forth by Congress. Akin to the systematically important financial institutions designation established by the Dodd-Frank Act,30 designating powerful online services as gatekeepers that merit dedicated scrutiny enables regulators to look at business practices not only in isolation, but also in terms of what systemic risks they may pose to the economy and the public interest.
Amid growing demand for government action to address online harms and increasing regulatory action abroad, the United States must urgently pursue aggressive antitrust action, updated competition policies, and robust federal privacy laws and rules. Looking to the future, a comprehensive new regulatory approach for online services is critical. While the regulatory framework presented in this report would be a significant undertaking, the cost of inaction would be much greater. A government that cannot understand, much less anticipate, the dangers and potential of new technologies will increasingly fail the public over the coming decades. With that in mind, this proposal lays out commonsense ideas to enable effective democratic regulation of online services now and into the future.
This report proceeds by outlining the harms from online services—illustrating the economic, consumer protection, privacy, and civil rights issues they raise—and identifying the gaps in the current regulatory landscape for addressing those harms. Next, it proposes a three-tier regulatory strategy before diving into each of the three proposed categories: online infrastructure, general online services, and gatekeepers. Each section outlines the target entities, regulatory logic, and new tools proposed. The authors then discuss the proposals in the context of how they relate to a cross-cutting digital policy issue—how to handle the locus of problems grouped as “harmful online content”—by outlining potential ways that expanded regulatory capacity could contribute solutions. Finally, the report discusses options for administration—whether by expanding powers at existing agencies, creating a new agency or body, or using some combination of these approaches—and closes with a note looking to the future.
The focus of this proposal is online services, which refers to services and products delivered through the internet. Importantly, for the purposes of this report, it excludes: 1) the core set of protocols, standards, and networking architecture that constitute the internet itself, and 2) simple, static content displayed using those protocols, which does not rise to the level of service provision. Beyond that, the scope is deliberately straightforward: If a service does not work without the internet, it is considered an online service.
In technical terms, this encompasses services offered in the application layer of a traditional internet stack but generally does not encompass telecommunications or networking infrastructure further down the stack, such as physical networking infrastructure or internet service providers. In other words, online services picks up where Federal Communications Commission (FCC) internet regulation currently leaves off, focusing instead on “edge providers”—defined by the FCC as “content, application, service, and device providers, because they generally operate at the edge rather than the core of the network.”31 This definition includes cloud services, modern operating systems, app stores, search engines, social networking services, multimedia sharing services, online communications services, digital advertising infrastructure, IoT products, software as a service (SaaS) products, algorithmic decision-making services, and online marketplaces of all kinds. It would not include basic internet protocols, core standards, or display-only content put online; for example, a garden variety HTML display website32 that lists and links text online would not be considered an online service. Though the vast scope of online services spans markets and regulatory areas, the commonalities are clear: Each is delivered over the internet, presents a regulatory gap, and faces common competition and consumer protection challenges that are endemic to online markets.
Conceptual models of the internet, including the internet protocol suite and the open systems interconnection model, use a “stack” framework to explain the operation of interconnected protocols that form the basis of the internet.33 Rhetorically, the idea of the stack is used to discuss the range of technologies and protocols from the hard networking architecture at the bottom to the most consumer-facing services at the top. Figure 1 illustrates a simplified internet protocol stack. Many of the online services discussed in this report fall within the application layer of a traditional stack, although the authors also include some services that span application and transport layer services. Within the application layer, user-facing websites are conceptualized as higher up in the stack compared with the unseen cloud services or content delivery networks that underlie them. While the interconnections between middleware, software, and web applications are complex, various, and changing, the distinction between services that are consumer-facing or those that are not can be meaningful from a regulatory standpoint.
Figure 1
Framing this universe as “online services” seeks to reflect the reality that, while the scope and variety of online services will only increase, not all businesses or market participants should be considered tech companies or digital platforms. Rather, the business practices employed in the provision of online services have common elements that can be regulated with a common logic, even as they span unrelated industries. Roughly, these common elements include:
The authors recognize the difficulty of balancing policy tensions and conflicting interests, as described in Ofcom’s 2019 report on regulating online services:
Online services pose particular challenges for regulators. This is because of their global nature, the fast pace of change, the complexity of online business models, the scale of online content and the variety of services available online. The links that may exist between different harms can create overlaps and tensions between policy aims. These are challenges for existing regulators, as well as for any future regulation of online services being considered by Government.37
While some online services markets are fast-moving or complex, others are stable or straightforward. The ubiquity of claims from industry—that their operations are too complicated to be appropriately regulated—presents an even stronger argument for an increase in the number of public servants who know the difference. Online services are not so complex as to be unregulatable, and opacity in their operations can be remedied.
There are meaningful definitional challenges to be worked out in this or other approaches.38 However, these challenges must be overcome if the United States is to functionally address the multitude of harms occurring from the unchecked power of both new technologies and dominant gatekeeper platforms.
The growth of the internet has produced significant social, cultural, and economic benefits for the United States. Providers of online services have helped shepherd the internet from its infancy into a more accessible digital layer that interweaves with most Americans’ lives on a daily basis. With the exponential growth of online services and their attendant benefits, however, a number of harms have also emerged, enacted or enabled by online services providers. While many Americans have grown up accepting these harms as a cost of engaging online, the harms generated by online services are not inevitable. Current problems are not necessary evils for the sake of digital innovation, and improved regulation has a dual role to play in promoting beneficial development and curbing predatory practices.
In order to support a vibrant, dynamic internet that serves the public interest, it is necessary to understand not just the benefits but also the harms from online services and the risks they pose to economic, social, and democratic health. These harms are salient and widespread, even where services are offered at low or no monetary cost to users—for example, free email or social networking platforms, which are subsidized by intensive data collection, online tracking, and targeted advertising. These harms tend to be disproportionately borne by marginalized groups—including people of color, low-wage workers, and women—whereas technology’s benefits asymmetrically accrue to more privileged groups.39 In aggregate, these issues amount to troubling threats to commerce, civil rights, and democratic function. To make these issues more legible to traditional regulatory approaches, they are grouped below into four overlapping and deeply interconnected areas: economic harms, privacy harms, consumer protection harms, and civil rights harms.
This survey of harms is necessarily incomplete. While a full examination of online harms is beyond the scope of this report, the limited information available also speaks to the profound asymmetry and lack of transparency in the online services space. This information asymmetry—the stark lack of data accessible to government and the public compared with the mountains of data held by digital platforms—is a persistent issue across different areas of harm. Indeed, the harms described below may only be the tip of the iceberg. Researchers are starved for data on online harms and competition, and many of these issues have only come to light through formal government inquiries, whistleblowing, or intrepid investigative journalism.40
The proliferation of the internet and digital communication technologies have produced new and complex online businesses. The largest of these businesses have developed communication and information services that have become essential to billions of consumers and are protected from new competitors by powerful barriers to entry. As noted above, these barriers exist because of inherent features of digital markets such as network effects, economies of scope and scale, data advantages, first-mover advantages, and other economic forces.41 They have been preserved, reinforced, and compounded over time by strategic acquisitions and successful efforts by firms to foreclose nascent competitors and discourage competitive threats,42 resulting in traditional problems arising from a lack of competition—higher prices, lower quality, and less innovation. In markets dominated by an incumbent digital gatekeeper, the threat of the dominant firm copying or killing any new innovations results in decreased investment, deterrence of entry, and decreased innovation in the digital platform industry.43 Big tech mergers likewise have adverse competitive effects on growing markets,44 and incumbent firms may acquire younger firms explicitly to curb innovation that threatens their position.45 More than 80 percent of Americans believe acquisitions from large online platforms are likely unfair and undermine competition.46 But with few alternatives, high switching costs—for example, the difficulty or inability to move personal data when shifting to a new service—and extremely powerful network effects mean that American consumers have a limited ability to “vote with their clicks.” Even with great effort, it is difficult to avoid using major firms; journalist Kashmir Hill described her experiment living without any services from five nearly inescapable technology companies as “hell.”47 This lack of choice further removes incentives for dominant players to innovate to improve services.
Centralization of research and development (R&D) resources at dominant firms may additionally result in selective or reduced innovation. A lack of external competition, for instance, discourages innovation,48 and internal research that threatens dominant business lines is often avoided, hidden, or systematically challenged.49 There may be significant opportunity costs to having only a few big U.S. technology companies driving the direction of technological progress for the economy more broadly, especially given the competitive incentives for dominant firms.50 Experts have also raised concerns about the national security risks of relying on only a handful of dominant global technology companies that may not prioritize U.S. national security interests and do not have sufficient competitive incentives to ensure continued innovation, performance, and efficiency.51 There is nothing wrong with firms pursuing innovations entirely compatible with their business models. But when such R&D capacity is concentrated among only a few major technology firms with similar incentives and limited demographic diversity, there is cause for concern about whether these innovations will benefit low-wage workers, address climate change, and benefit the public interest, or whether they will continue to concentrate America’s R&D efforts around issues such as selling online advertising.52
A lack of competition also produces pricing harms, even when the direct consumer price is zero or the upfront consumer price is competitive. Indeed, a multisided platform—for example, a website that brings together consumers, business users, and advertisers and provides a platform for sales and interaction—may charge fees to business users that are directly passed on to consumers down the line. Consumers who enjoy low prices from digital giants today might also face higher prices in the future: Incumbent firms may use price-cutting strategies or subsidies to kill potential competitors and build or maintain market power, producing lower prices in the near term but ultimately resulting in higher prices and lower quality. Amazon, for example, dropped its diaper prices by more than 30 percent, effectively curbing the growth of online retailer and undercutting the company whenever it dropped prices.53 Through cross-subsidization with other business lines, Amazon was able to absorb losses on baby products in the short term, “no matter what the cost,”54 in order to maintain its dominant market position and price-setting ability in the long term, catalyzing the forced sale of the once-burgeoning diaper retailer. In her landmark paper “Amazon’s Antitrust Paradox,” Lina Khan argued, “The fact that Amazon has been willing to forego profits for growth undercuts a central premise of contemporary predatory pricing doctrine, which assumes that predation is irrational precisely because firms prioritize profits over growth.”55 While many markets experience this kind of short-term corporatism, its effects in digital markets are more harmful. As noted earlier in this report, digital markets are prone to tipping, wherein one firm is likely to “win” and maintain most of the market after gaining an early lead. Firms then leverage existing dominance for further expansion, tipping, and entrenchment in adjacent markets, making the economic consequences of unchecked predatory behavior particularly high.
American small businesses are likewise harmed by a lack of competition among digital platforms. Facing few alternative choices, high switching costs, and little power to change platform conditions, American small businesses face a high degree of platform precarity: increased risk due to heavy reliance on a handful of dominant platform services over which they have little influence or recourse if problems arise, even when platforms are treating them unfairly. Dominant platforms use this knowledge to extract rent in the form of unfavorable pricing, terms, agreements, and more;56 examples include third-party business users such as restaurants using food delivery apps, third-party retailers creating storefronts on major online retail services, and content creators monetizing their video or audio content. Small and medium-sized businesses are forced to invest significant resources to compete effectively on online platforms, but sudden, unilateral changes in terms,57 ranking,58 pricing,59 design, or a sudden suspension60 can wipe out the value of a business’ investment. Even getting out of these service arrangements can be costly: Cloud services, for example, may make it cheap to transfer data into the service but charge ultra-high rates for egress fees to leave a service.61 Worse, platforms may exploit data about a business’ sales and products to develop copycat products and undercut small businesses,62 potentially even self-preferencing first-party services through pricing, data, design, ranking, and bundling strategies.63 Increasingly, dominant platforms’ theft of content threatens internet openness and undermines small or growing firms.
American workers are also harmed under the status quo. When dominant firms drive out competitors and achieve market capture, firms become labor monopsonists,64 meaning that they acquire disproportionate power to set and decrease wages because they face little competition that might otherwise motivate a competitive wage and safe working conditions.65 Worker abuse is easy to disguise through the ubiquitous use of opaque business software and algorithmic management systems, which may rely on surveillance to monitor and shape worker behavior.66 While some of these issues can be addressed through updating and robustly enforcing labor laws, the competitive failings of digital markets will continually put downward pressure on wages and working conditions in monopsonized labor markets.
A persistent lack of transparency and data asymmetry exacerbate these problems. While workers or business users may feel that abuse is occurring, it is difficult to investigate problems without greater data access. These issues are of growing importance to Americans, with 81 percent of voters saying they are “concerned about consolidation among Big Tech corporations hurting small businesses and consumers.”67
Historically, while businesses such as telephone networks have also been protected by strong network effects and high barriers to entry, online service providers are unique in that many also surveil their consumers—sometimes without consumers’ awareness—and use the information they gather to manipulate user behavior to increase usage and revenue. Other companies then buy this information from surveillant firms, develop predictive statistical models, and sell those models for wider use. The application of these models materially affects peoples’ lives in ways that are often hidden from them; the resulting invasions of privacy and invisible impacts on people’s health, economic prospects, education, and liberty have produced novel forms of harm to society. Due to the complex and sometimes deliberately obscured workings of online services, it can be difficult or impossible for individuals to understand, address, or even identify the origin of these harms—let alone choose a better option if one is available.
Unwanted and invasive data collection, processing, and sale have become standard practice in online services industries, and Americans are overwhelmingly concerned about the data platforms hold.68 The scope and detail of corporate data collection and consumer surveillance are astounding. For example, Google reportedly has acquired information on 70 percent of all U.S. credit and debit card transactions69 to combine with its detailed user profiles. An entire industry has grown around creating and selling constant, unwanted records of billions of people’s locations at scale in gross detail:70 One analysis found that location trackers in common, innocuous mobile phone apps were updated more than 14,000 times per day, identifying individuals’ location down to within only a few yards.71 These data are profoundly abused. Companies collect consumer contact information and movements without consent;72 perpetuate the pretense that consumers give informed consent with the click of “I agree”;73 use deceptive disclosures and settings to trick consumers into allowing data sharing with third parties;74 track consumers’ location within a few feet inside their homes;75 track consumers’ location even after tracking is turned off;76 develop new products using consumers’ personal emails, photographs, and conversations;77 track people’s ovulation data without consent;78 and then too frequently fail to secure the massive troves of intimate and valuable data they acquire. It is not just dominant firms engaging in this behavior: In some cases, small businesses and third-party data buyers are the worst abusers of consumer privacy.79
Indeed, privacy harms are acute in combination with competitive harms. Experts have shown that firms that achieve market dominance and successfully suppress competitive threats are able to lower privacy protections in order to pursue and extract greater data gains from consumers.80 Consumers, without a reasonable choice of substitutes, are forced to put up with suboptimal privacy protections and even privacy invasions. Within digital markets, experts including Howard Shelanski have argued that “one measure of a platform’s market power is the extent to which it can engage in (data usage that consumers dislike) without some benefit to consumers that offsets their reduced privacy and still retain users.”81 As illustrated by Dina Srinivasan, Facebook’s pivot away from privacy protection toward privacy exploitation upon achieving monopoly status is emblematic of this power, with consumer data extraction constituting part of the firm’s “monopoly rent.”82
The collective costs of individual privacy incursions, of which consumers are often unaware, are staggering. These costs are not just economic—although billions of dollars have been lost through corporate negligence to protect these data, especially sensitive information concerning individuals’ credit, finances, and identity83—but also democratic, social, and humanitarian. Troublingly, Americans have changed their social and political behavior because they know they are being watched by corporations and law enforcement.84 Ambient surveillance has chilling effects on expression, civil liberties, and freedom of movement, particularly for Black and Hispanic communities that are persistently oversurveilled and overpoliced.85 Americans’ personal interactions, behavior, and political activity have become commodities to be tracked without consent, bought, and sold. As companies reach beyond merely advertising to manipulating people’s behavior,86 the societal costs and implications are profound.
Consumer protection issues in online services include but extend beyond traditional privacy concerns:87 Issues with fraud, scams, manipulation, discrimination, and systemic failures in content promotion and moderation have leveled devastating individual and collective harms.
A scale-at-any-cost growth mindset,88 overly broad interpretations of intermediary liability laws that cover the sale of physical goods,89 and other factors have disincentivized the development of more reasonable responsibility for consumer protection. For years, lawmakers have asked e-commerce sites to stop selling unsafe, banned, fraudulent, or knock-off products and asked other websites to stop advertising them.90 A lack of quality control makes it easy to place false listings or reviews online to scam consumers, scam businesses, damage competitors, harass victims, and divert traffic from legitimate small businesses.91 Negligent safety standards on large platforms have enabled bad actors to commit elaborate frauds, ranging from digital advertising schemes that scam advertisers to fake accommodations listings that defraud would-be guests to marketplaces that fail to protect users from scammers at scale.92 In some cases, the gap between self-defined platform terms and actual enforcement across these issues is apparent.93
Due in part to the shift to online services during the pandemic, people are facing growing threats from long-standing consumer protection and cybersecurity issues. Losses to identify fraud, for example, topped $56 billion in 2020.94 These costs are disproportionately felt: One analysis found that “Black people, Indigenous people, and People of Color (BIPOC) are more likely to have their identities stolen than White people (21 percent compared to 15 percent), and BIPOC people are the least likely to avoid any financial impact due to cybercrime (47 percent compared to 59 percent of all respondents).”95
Beyond sensitive financial and identity issues, the unprecedented amount of detailed behavioral data held by online services firms also poses unique consumer protection challenges. Platforms are able to exploit behavioral shortcomings and biases among consumers in real time to a greater degree than previously feasible.96 They may intentionally complicate the process of changing privacy settings, opting out of data collection, deleting accounts, canceling services, and more.97 These designs may hide or misrepresent costs,98 fee structures,99 and data collection.100 In a digital environment, firms are able to more fully manipulate the buyer experience, making consumer manipulation of heightened concern.101 Some firms employ deceptive behavioral design, sometimes called “dark patterns,” which have been found to successfully manipulate consumers into giving up time, money, or information.102 The ability to use detailed data and pricing systems has given rise to new forms of dynamic pricing, which too often replicate long-standing biases against historically marginalized communities.103 Nearly three-quarters of Americans think this type of personal data-driven dynamic pricing is a “major or moderate problem.”104
Online services have also given abusers and harassers more ways to locate and target victims while regularly failing to provide people with sufficient tools for preventing, curbing, or avoiding those attacks.105 A recent poll found, “Of the types of harms people experience online, Americans most frequently cite being called offensive names (44 percent). More than 1 in 3 (35 percent) say someone has tried to purposefully embarrass them online, 18 percent have been physically threatened, and 15 percent have been sexually harassed.”106 Numerous online service companies have failed to take adequate steps to prevent these harms from occurring.107 Over the past two years, the number of teenagers who reported encountering racist or homophobic material online almost doubled.108 Marginalized communities—especially transgender people, immigrants, people of faith, people of color, and women of color—are disproportionately harmed through negligent or actively harmful platform business models around content and bear the brunt of their collective costs.109
Online services regularly introduce risks to Americans’ civil rights and liberties.110 Use of digital technologies—including software, algorithmic decision-making systems, digital advertising tools, surveillance tools, wearable technology, biometric technology, and more—have introduced new vectors to continue the deeply rooted historical exploitation of and discrimination against protected classes. Because privacy rights are also civil rights, these harms are inextricably linked to the privacy harms described above, wherein mined data feed into algorithms that are used to profile individuals, make decisions, target ads and content, and ultimately lead to discrimination.111
Leading scholars and advocates have exposed the numerous risks that automated decision-making systems—encompassing everything from static algorithms to machine learning to AI programs—pose to civil and human rights.112 These systems can produce deeply inequitable outcomes, including and beyond issues of algorithmic bias.113 Discrimination can occur at any point in the development process or produce, obfuscate, and launder discriminatory use. Already, they have resulted in a slew of civil rights violations that materially affect Americans’ liberty, opportunity, and prospects. Algorithmic decision-making systems have produced and reproduced discrimination in recruiting,114 employment,115 finance,116 credit,117 housing,118 K-12 and higher education,119 policing,120 probation,121 and health care,122 as well as the promotion of services through digital advertising123 and beyond.124 Algorithmic racism in particular extends the project of white supremacy in pernicious ways:125 With a glut of consumer data and the veneer of technical objectivity, online services companies have myriad ways to discriminate among consumers and obfuscate that discrimination.126 For instance, digital advertisers can use proxy metrics to enable discrimination in advertising without technically using protected classes,127 although Facebook has been sued for allowing discrimination based on protected classes explicitly.128 Insurance, credit, and financial companies can bake historical data, which reflect long-standing inequities and biases, into decision-making algorithms that enable them to reproduce systemic racism and other biases while using a seemingly “objective” algorithm that processes applications in an identical manner—churning out preferential products and opportunities for white, wealthy people as they have for decades.129
Technology-enabled discrimination is especially dangerous because the application of these tools can be hidden and nonconsensual, limited forms of redress exist, and technical processes are often wrongly assumed to be objective, thereby receiving inappropriate deference or insufficient scrutiny. New AI and algorithmic hiring tools, for example, have been hailed for their “efficiencies,” yet are found to compound existing issues in disability-based discrimination, despite long-standing Americans with Disabilities Act protections.130 A range of algorithmic and platform design choices can likewise enable discrimination.131
Facial recognition and other biometric surveillance technologies erode civil liberties, particularly for communities of color.132 The biases in these technologies133 and their use by law enforcement134 have led to traumatic violations of civil liberties, including a number of recent wrongful arrests of innocent Americans who were misidentified by faulty facial recognition software.135 But more broadly, their increasing use in public spaces and employment as tools to continue the overpolicing and oversurveillance of people of color threatens civil liberties, chills political speech, and inhibits freedom of movement and assembly.
Content moderation challenges and negligence also introduce asymmetric risks to protected classes. Platforms’ failures to prevent the exploitation of social networking for purposes of harassment, discrimination, hate speech, voter suppression, and racialized disinformation have made long-standing problems newly urgent. Furthermore, major platforms have been found to increase radicalization and participation in extremist groups.136 At the individual level, these problems have subjected people to harm and serious duress137 and enabled the deprivation of rights, including the right to vote.138 Civil rights experts have drawn parallels between the discriminatory nature of these business decisions and platform designs and the public accommodation laws that protect against discriminatory practices in brick-and-mortar businesses, highlighting the need to update and reinforce current digital protections.139
Collectively, the sheer quantity and amplification of such civil rights-suppressing content introduces barriers to and discourages full participation in public life and cultural discourse by already excluded groups. The prevalence of false information and propaganda on social media in particular can grossly warp public discourse and societal understanding of public events. Misinformation has been used to maintain and advance racist, sexist, transphobic, and other prejudices, while “astroturfing” strategies—wherein coordinated networks of accounts, including “fake” accounts not representing “real” people, artificially inflate the popularity and visibility of certain posts—are used to misrepresent the prevalence of these attitudes. For example, despite the majority of Americans supporting Black Lives Matter, 70 percent of Facebook posts from users discussing the topic in June 2020 were critical of the movement.140
Beyond posing risks to specific enumerated rights and liberties for protected classes, online services have reified, maintained, and extended racism, sexism, and other social prejudices generally in the United States, through both their technology development and business model negligence. For example, Dr. Safiya Noble’s pioneering work illustrated that, for years, searching “black girls” on Google returned pornographic search results and ads, whereas searches for “white girls” did not.141 Similarly, searches of Black-identifying names disproportionately returned ads mentioning “arrests” compared with searches of white-identifying names.142 Numerous other instances of search engine and predictive text results enhancing and extending social discrimination abound,143 and similar problems exist in voice technologies, facial recognition, and other biometric and visual processing techniques.144
Across these four overlapping and interconnected areas of harm—economic, privacy, consumer protection, and civil rights—the information asymmetry and power of online services firms threaten to impede understanding and responsible regulatory solutions. Specifically, massive spending power and political leverage strongly influence regulatory and political environments, as well as firms’ ability to shape media and public discourse in their favor. Finally, as noted above, the lack of transparency regarding economic activity, data collection, and content moderation makes it difficult to identify or verify suspected harms.
Finally, Americans have long recognized the unique political power of media industries and the importance of pluralism and diversity in the press. Every major emergent communications technology in modern history, from the printing press to television, has engendered new challenges and problems.145 However, Americans are especially concerned about the power online services have over public discourse and the political system. Recent surveys show that approximately 81 percent of U.S. adults, and majorities of voters of both political parties, believe that technology and social media companies “have too much power and influence in politics.”146 Likewise, 77 percent of American adults believe it is a major problem that online search and social media platforms control what people see on their platforms. Simply put, concentrated power in online services—particularly among social media, search engines, and cloud infrastructure—are cause for democratic concern and action.
Existing laws, authorities, and agencies can address a subset of interlocking online services harms outlined above. In particular, the Center for American Progress strongly supports more aggressive antitrust action,147 more robust competition policies,148 increased privacy and civil rights capacity at the FTC,149 and strong federal privacy legislation or rules.150 The 117th Congress has an opportunity to achieve key gains in these areas by passing the tech antitrust package—which has been reported favorably out of the House Judiciary Committee151 and for which companion bills have been introduced in the Senate152—fully resourcing the DOJ and FTC,153 and settling on a strong federal privacy proposal. Significant progress is on the table.
Looking ahead, however, even an optimistic reading of these proposed updates shows gaps would persist in the government’s ability to tackle the vast scope of online services harms in a timely and effective manner. As outlined below, existing systems of regulatory oversight are primarily reactive: Judicial scrutiny and dedicated, but often narrower, piecemeal legislation have struggled to keep pace with technological and market change. In a vacuum of regulatory scrutiny, consumer harms have accumulated, predatory practices have become industry standards, and dominant players have entrenched and expanded their holdings. Over time, a regulatory “debt” has built up where existing statutes and sector-specific regulations have not been sufficiently updated or applied to novel problems. Labor laws, for example, have lagged behind developments in algorithmic workplace management systems. Effective regulatory oversight must grapple with not only emerging issues but also the regulatory debt that has developed over past decades.
Historically, developing remedies has taken years, sometimes more than a decade, to reach resolution after harm has occurred. While it is certainly possible for congressional oversight to dedicate the required expertise to online services regulation—as seen in the historic 2020 House antitrust report154 and resulting bipartisan legislative proposals from the House and Senate in the 117th Congress—it is impractical for them to do so for dozens of different online services industries presenting novel problems, or old problems in new bottles. This is particularly true for small and medium-sized players that require regulation but lack the public recognition or political attention merited by digital gatekeepers. Significant federal investment in public interest oversight and administrative bodies is needed to understand and rectify the problems that have proliferated during the past 20 years.
Basic regulatory capacity has not kept pace with the growth of online services. This section surveys current regulatory tools and identifies the outstanding gaps. It presents a mix of current gaps and those that would likely remain if privacy and competition developments are enacted. 
Economic harms could be partially addressed through existing antitrust laws, including the Sherman Antitrust Act and Clayton Antitrust Act,155 as enforced by the U.S. Department of Justice (DOJ), FTC, and state attorneys general.
However, over recent decades, a successful movement to narrow the application of antitrust laws to a limited consumer welfare standard has allowed monopolies to flourish across industries. The anemic antitrust enforcement that has resulted has enabled increased concentration of power in many sectors, including technology and online services markets.156 Existing authorities are limited in their abilities to increase competitive pressure on already dominant firms. Furthermore, limitations exist in addressing market dominance arising from inherent network effects; conventional antitrust does not necessarily forbid monopoly in the absence of exclusionary, improper, or predatory acts. Where applicable, antitrust tools can be slow: With important exceptions, such as merger reviews, many are limited to after-the-fact intervention. These qualities have hampered antitrust effectiveness in the online services space, where remedies are sometimes pursued too late.
Revived enforcement is essential to remedying current problems and promoting competition.157 Recent actions from U.S. enforcement agencies—such as the antitrust suits filed by the FTC, DOJ, and state attorneys general against Google and Facebook, as well as other ongoing investigations—are positive steps toward these goals.158 These cases use existing authorities to address economic harms including negative impacts on innovation, pricing, and labor.
Unfortunately, complex court cases have yearslong timelines. The outcomes of these cases are uncertain—and even more so due to the difficulties in applying a deficient consumer welfare standard to digital markets. The conservative shift in the federal judiciary over past decades and the high bar set by existing antitrust laws and court decisions further complicate enforcement. Even in the event of a successful case, the ensuing appeals process may take years, and selected remedies may fall short of those that would most effectively address anticompetitive effects, such as structural separation, reversal of mergers, or divestiture.
During the years it takes for cases to work through the courts, tech giants will continue to expand, entrench, and potentially abuse their dominance. Monopolists under scrutiny may well outlast any viable competitors and the government administrations that bring suits to challenge them in the first place. Injunctive relief during investigations may help prevent further consolidation but is only a temporary, limited measure. Some experts have argued that the threat of potential antitrust action makes dominant companies operate more cautiously and accept competitive measures that they would otherwise oppose. However, companies often calculate the bare minimum required159 to escape with their dominant market share intact;160 given the scope, scale, and importance of online services to the United States, self-regulation and deterrence are no longer viable strategies.
Beyond revived antitrust action, a number of complementary competition policy reforms are needed.161 These include increasing the focus on anticompetitive conduct by dominant firms, making antitrust enforcement more transparent and robust, and setting out clear new rules against self-preferencing or discrimination by dominant platforms.162 In 2020, the House Judiciary Subcommittee on Antitrust, Commercial, and Administrative Law issued a landmark report on digital markets and gatekeepers, illuminating a number of the anticompetitive issues in digital markets.163 In 2021, the House Judiciary Committee reported favorably out of committee a suite of new antitrust and competition policy bills designed to address these issues.164 For covered online platforms, the bill introduce pro-competitive provisions around nondiscrimination and self-preferencing,165 mergers and acquisitions of competitors,166 interoperability,167 merger filing fees,168 lowering barriers to state antitrust enforcement,169 and eliminating fundamental conflicts of interests between platform operators who wish to compete within their own marketplace.170 The American Innovation and Choice Online Act would also call for the creation of a new Bureau of Digital Markets within the FTC to handle the increased workload and specialization that digital markets require.171 These proposals, and the Senate companion bills,172 offer a powerful opportunity to address some of the most pernicious anti-competitive practices used by online services gatekeepers today. Looking to the future, additional work will be needed as new markets, predatory practices, and gatekeepers emerge—including and beyond those that would currently be covered under the bills’ proposed tests.
A patchwork of state and federal laws target consumer protection issues posed by online services, from the Better Online Ticket Sales Act,173 to the Children’s Online Privacy Protection Act,174 to the Restore Online Shoppers’ Confidence Act,175 among others. Broadly, however, the FTC is charged with protecting consumers by stopping unfair, deceptive, and fraudulent practices—which includes practices employed by online services providers.176 The FTC is generally the primary agency addressing consumer protection issues from online services, though other agencies such as the Consumer Product Safety Commission and Consumer Financial Protection Bureau (CFPB) may play a role depending on the specific sector or focus of a company. State attorneys general also have consumer protection responsibilities, although many states have weak or ineffective unfair and deceptive practices laws.177
On a federal level, years of aggressive opposition have limited consumer protection and the FTC’s ability to effectively enforce them. Scaled-back use of FTC rule-making in recent years has likewise contributed to growing deficiencies in consumer protection online. Some observers, including former FTC commissioner and current CFPB Director Rohit Chopra, have argued that the agency has been shy about using the full range of existing authorities.178 Others have suggested that it is “overburdened” with an immense jurisdiction and limited resources, pointing to a history of companies requesting to be placed under FTC oversight in order to “get its issues lost amid the issues of other companies.”179
Despite its broad consumer protection and competition mandates, the FTC is not a large agency. It fulfills its mission with limited capacity: about 1,160 full-time employees to cover consumer protection in most sectors. This number has dropped significantly over the past several decades. Consumer Reports recently noted that, since 1979, “the economy has grown nearly three times while the FTC’s capacity has decreased 37 percent.”180 The FTC Office of Technology Research and Investigation has only a handful of staffers to support work across the agency.181
Enhanced consumer protection regulation and new legislation are required to protect Americans online. Given the scale and variety of consumer protection harms from online services, existing FTC authorities and capacity are manifestly insufficient. However, as noted below, recent developments are a promising start to restoring and elevating the agency to the capacity and authority needed to fulfill its mission. 
The United States is unique among its peers in that it still does have a national data privacy law. Consequently, it also lacks a designated data protection agency. In lieu of such a body, the FTC serves as the de facto data privacy agency: Its authority over unfair and deceptive acts or practices has been brought to bear against online privacy and security violations,182 and it serves as the regulator for the 1998 Children’s Online Privacy Protection Act. As FTC Commissioner Rebecca Kelly Slaughter noted in 2020, however, “In enforcing data privacy, the Commission does not have the most straightforward tools. The FTC has done an impressive job of attempting to curb the worst abuses in this space without the benefit of a federal privacy law, civil penalty authority, or anywhere near the dollars or the bodies that other countries devote to data privacy protection.”183 Indeed, the FTC has only 40 employees focused on data protection.184 By comparison, the Irish Data Protection Commission, the primary EU regulator for many large U.S. tech platforms, is only one of the EU’s 31 data protection supervisors185 and has three times as many staffers.186 It still faces regular criticism for its slow pace and large workload.187
Absent comprehensive federal protections, states have also increasingly adopted privacy protections through legislation or statewide referendums, starting with California in 2018 and 2020188 and Virginia and Colorado in 2021.189 Given California’s influential status as a population center and the home base to many technology companies, its state laws have unique nationalizing effects in the absence of federal law—for example, the 2003 California Online Privacy Protection Act was the first state law to require privacy policies to be posted on a website.190
Thus far, the limited powers of existing regulatory agencies have not sufficiently protected and empowered Americans in the current data environment.191 For example, a recent New America Open Technology Institute report noted:
The FTC’s approach primarily relies on corporate self-regulation under the notice and consent model, bringing enforcement actions against companies that have deceptively violated their own privacy policies and public representations made to users about how they protect their privacy and security. While this strategy has led to a number of enforcement actions over the years, the notice and consent model relies on a number of faulty assumptions, including the notion that the average user can meaningfully consent to privacy policies.192
However, as outlined in a recent report by the Electronic Privacy Information Center,193 the FTC does have unused authorities it could exercise around online privacy. Commissioners have shown recent interest in reviving the FTC’s latent authorities to tackle data privacy, including the Magnuson-Moss Warranty Act rule-making authority.194 Under new leadership, there are indications that the FTC may move to make better use of these authorities,195 including over privacy harms.196 And given the pressing need for consumer protection online, lawmakers have recently proposed increasing funding to the FTC to expand its privacy focus.197 As noted above, these efforts must overcome the FTC’s capacity constraints, limitations to its traditional rule-making ability, and reliance on consent decrees—which have become increasingly routine and symbolic for corporate America. Privacy must also compete with other issues at the FTC, whose limited capacity requires picking and choosing issues within its broad mandate.
Complementary to any renewed privacy efforts at the FTC, strong federal privacy legislation is necessary and overdue. Fundamentally predatory data collection practices must be prohibited. The federal government must establish increased enforcement capacity to guard against the numerous harms of nonconsensual data collection, sale, and discriminatory use. Federal legislation should include robust civil rights protections, strict limits on the use of personal data, limitations on consent-based models, enhanced individual rights and privileges, and action paths to defend new rights for individuals and state governments. Recognizing a critical need for increased capacity, federal privacy proposals from both Republicans and Democrats include giving the FTC increased rule-making authority over privacy,198 and Sen. Kirsten Gillibrand (D-NY) has proposed creating a new data protection agency.199 As part of a new social spending package, the U.S. House of Representatives also proposed significant investments in the FTC to create a new privacy bureau.200
These steps would provide a badly needed foundation for guarding Americans’ privacy. A new privacy law, however, may not be able to address new and creative abusive behaviors that will inevitably arise. Few proposals would holistically grapple with the chilling effects of pervasive, ambient personal and biometric surveillance. The practice of surveillance advertising201 may be difficult to curb so long as the incentives stay in place and the behavior remains legal. Finally, privacy is not the sole lens through which to judge the impact of online services; tensions must be dynamically balanced among privacy, security, competition, transparency, and other priorities.
Americans across the political spectrum are strongly in favor of robust federal privacy protections,202 and in order to properly enforce those restrictions, regulatory bodies need additional administrative rules and capacity along those lines. Here, as with competition and antitrust approaches, new statutes, enhanced capacity, and specialist oversight are required to effectively govern online services. 
The evolution of online services has outpaced the application and interpretation of civil rights laws to digital properties and transactions. In general, civil rights laws apply broadly, including to online behavior, transactions, and properties; the law, for example, does not distinguish between discrimination in employment advertisements in a newspaper and online. Yet in practice, the protection of civil rights online has lagged behind emerging and present risks, outpacing the ability of the DOJ, Equal Employment Opportunity Commission, U.S. Department of Health and Human Services, U.S. Department of Agriculture, and other federal and state enforcement bodies and officers to identify and investigate potential violations. Numerous groups, including the Center for American Progress, have called for the creation of an Office of Civil Rights at the FTC to strengthen its ability to prevent discrimination and protect equal opportunity online.203
However, beyond capacity, some argue that a lack of clear case law or uncertainty around what case law is applicable to novel technologies can impede enforcement efforts. Where relevant case law does exist, application of the existing doctrine to online services discrimination cases is not always straightforward.204 Significant work is required to grapple with the substantive challenges posed in some, though certainly not all, online or data-driven discrimination cases where the structures or business systems involved in the online service provision do not map neatly onto existing templates in case law. Once again, the numerous barriers that exist in algorithmic accountability and transparency205 have further complicated effective civil rights enforcement.206
While in most cases there is no question that existing civil rights laws apply online, any approach to regulating online services should forcefully reject certain questions about how existing laws apply, ensure that there is effective and robust enforcement of these laws online, and close any loopholes that may exist for narrow “frontier” areas where existing laws may not have predicted harms. Given the sheer scope of the new transactions, interactions, and digital properties involved in the provision of online services, significant, proactive work can ensure that Americans’ civil rights are protected and prioritized.
Consider the application of existing tools to the case of Facebook, Inc. (The company recently changed its corporate name to Meta Platforms, Inc., but for the purposes of this paper, will continue to be referred to as Facebook.207) Facebook appears to have acted anti-competitively.208 In 2021, the antitrust case brought by the FTC against Facebook209 was dismissed by a federal judge210 and refiled by the agency.211 If the case is successful and the FTC is granted all requested remedies—including rolling back Facebook’s acquisitions of WhatsApp and Instagram—the divestitures would remedy ground gained anti-competitively. But they would only go so far in preventing abuse of consumers on Facebook’s main platform, Facebook Blue: the largest social network in the world.212 Moreover, Instagram would be in a position to copy Facebook’s problematic practices around privacy, discrimination, competition, and general lack of accountability for design choices and business decisions. Breaking up a predatory company may reduce the scale of harm and introduce new competitive incentives, but it does not necessarily prevent continued predatory practices—described by Harold Feld as a “starfish problem.”213 Alternatively, the proposed tech antitrust bills would address a range of issues—including ending self-preferencing, encouraging interoperability, and instituting a higher bar for future efforts to acquire nascent competitors. However, since the newly independent Instagram would likely not meet the bill’s requirements for a covered platform, it would not, unlike Facebook, be subject to the new competition policies—even if some of these practices should arguably not be allowed at all. Indeed, new data portability might allow Instagram to implement a nearly identical version of Facebook’s targeting algorithms—likely faster than other competitors due to familiarity and existing technical architecture—allowing it to crush new entrants. A new federal privacy law may return greater control to users in managing the huge amount of data that Facebook and Instagram have collected on them but would still allow for much of the first-party ad targeting that brings in a majority of the platforms’ revenue. Similarly, outright banning of targeted advertising would leave a significant advantage to Facebook, which is in a prime position to pivot to contextual advertising through its first-party properties.214 Finally, while Facebook has previously settled lawsuits around discriminatory advertising, the public will still have few assurances that the company or divested businesses do not continue to harm civil rights.215 Aggressive antitrust action and structural remedies, new competition laws, and new privacy laws are all clearly needed, but would still leave gaps that allow for consumer, competitive, and civil rights abuses perpetrated by Facebook and Instagram.
Even in best-case scenarios for critical competition and privacy updates, significant gaps would remain in the U.S. government’s ability to anticipate and remedy online services harms. To effectively govern online services, U.S. regulators need to be empowered with proactive rule-making abilities that can curb problems before or as they occur. Such proactive rule-making powers—sometimes called “ex ante” regulation—are distinct from reactive or “ex post” approaches, which are litigated after harms have occurred. Proactive rule-making could identify and prohibit harmful measures prior to significant harm or as harms are occurring.216 In other words, this report proposes complementing after-the-fact antitrust enforcement by adding new restrictions and regulations that help prevent harm across multiple areas.
Regulatory tools go hand in hand with antitrust enforcement, which is the government’s best vehicle to address persistent anti-competitive behavior by individual companies. But, as noted above, antitrust tools are not necessarily well-matched to the range of noncompetitive harms posed by online services of all sizes. Yearslong timelines and uncertain outcomes further underscore the need for a parallel regulatory process. Going forward, adding new oversight powers could more quickly surface or address problems that require antitrust action. A center of excellence within the executive branch could expand the courts’ enforcement and oversight options when determining appropriate remedies. As FTC Commissioner Rebecca Kelly Slaughter testified before Congress in March 2021, “Effective enforcement is a complement, not an alternative, to thoughtful regulation. That is especially true for regulatory models that cannot be effectuated by ex post enforcement actions, even those with the broadest deterrent effect.”217
Increased regulatory oversight of online services can be a sensible complement to existing sector-specific regulations. In taking a cross-cutting approach, regulators could address the common aspects of problems that are unique to the online nature of service provision without necessarily affecting the many different industry-specific rules developed to govern the underlying services themselves. This is particularly true for sector-specific regulations in transportation, finance, labor, and other areas where there may be overlap. In The Case for the Digital Platform Act, Harold Feld writes that much like the pharmacy within a grocery store is regulated without subjecting the entire grocery store to pharmacy constraints, the online services offered by that grocery store can have dedicated regulation without necessarily subjecting its various other business practices to identical rules.218 Consider also, for example, how the U.S. Food and Drug Administration regulates prescription medication production very differently than it does growing vegetables—and within that, regulates industrial-scale vegetable production with some precision but stays out of home gardening. In the same way, dedicated online services regulation can scale by risk level and industrial heft where needed, rather than subjecting major e-commerce platforms and hobby webhosts to identical rules.
To be clear, while the proposals here are designed for online services, they are not all designed to apply to every online service; rather, online services are the universe in which common problems arise, and the rules proposed here would enable regulators to target specific problems within that universe.
This report advances a three-part regulatory framework for online services. It groups the two most distinct subsets—infrastructural services and dominant gatekeepers—apart from online services more generally. Services can opt in to online infrastructure regulations, subject to certain restrictions, qualifications, and regulator approval. The services that do not self-select as online infrastructure will fall into the general online services tier, regardless of their size. Those qualifying as gatekeepers will face additional, targeted regulations beyond the general online services regulations. The following sections sketch the parameters of each tier, noting target services, proposed regulation, and interactions with other tiers.
Online services provide essential infrastructure for the American economy, culture, and society. Cloud infrastructure, content delivery networks (CDNs), web hosts, and data analytics services are the quiet but dynamic backbone of the commercial internet. These online services now act as infrastructural components to other economic and social activity. They are generally lower on the stack than the more consumer-facing services at the top of the stack. (For more on the conceptual model of the internet stack, see Figure 1.) They are constant and ubiquitous, even if invisible to most internet users, and for this reason are often overlooked by regulators and the public until something goes wrong.219 A special focus on online infrastructural services is warranted due to their tremendous impact on the economy, the environment, cybersecurity, human rights, and freedom of expression.
To preserve, secure, and strengthen online infrastructure, this report proposes a mix of public interest obligations—such as common carriage, interoperability, security, and environmental protections—with dedicated intermediary liability protections independent of those under Section 230. The proposal targets largely unregulated online infrastructural services that are higher up the stack from FCC control of internet service providers at what are sometimes called “edge providers.”220 This approach adopts what scholar Annemarie Bridy describes as “layer-conscious internet regulation,”221 and adds to the discussion a distinction between infrastructural and other providers within the application layer, not just between the application and network layers.
For now, this proposal stops short of suggesting that these services be treated as public utilities. There are myriad examples of common carriage principles being applied to services that are neither public utilities nor regulated monopolies. The authors’ intention is that no lawful actor wishing to pay the established rate for online infrastructural services shall be denied or otherwise discriminated against so long as they are not publishing illegal content.
The rationale for this proposed mix of obligations and protections is three-fold: (1) discriminatory pricing practices can stifle innovation and competition; (2) infrastructural digital services pose different regulatory problems than the consumer-facing websites, services, and apps they support; and (3) there is substantial risk to freedom of expression when content regulation is pushed lower on the stack.
First, equal access to infrastructural services is foundational to ensuring free and fair competition across all online services that depend on them. In cases where these services are provided by companies offering other online services, there may be strong business incentives to opaquely employ differential pricing or restrictions in access. Additionally, this is a market where companies have erected significant and asymmetric barriers to customer entry and customer exit—for example, by levying much higher costs to transfer data out of a service than to import it.222
Second, the business decisions of online infrastructure have outsize effects on the security, accessibility, and resiliency of online services. They also have a significant material impact on energy consumption and the physical environment.223 Towns in which data centers are located face difficult choices around dedicating the water and electricity resources required to support them; a typical data center uses as much water per day as a small city, and many are located in arid, drought-plagued regions of the country.224 And while infrastructural services have been designed for performance, security, and reliability, they are generally not structured to offer the transparency and due process provisions that responsible content management requires. Forcing greater public visibility and monitoring systems into these services may adversely affect efficient, secure transmission of internet services to businesses and consumers. Taken together, this presents a strong argument for pushing content regulation to the edges—where consumer-facing services can more reasonably be expected to perform rights-respecting content moderation225—while also developing dedicated regulation for the security, environmental, and other aspects that are specific to online infrastructure.
Third, decisions by online infrastructural services have powerful effects on freedom of expression.226 Lacking due process, transparency, or accountability, infrastructure-level moderation offers blunt choices that, as Jack Balkin notes, may have significant collateral effects.227 Restrictions on arbitrary content-based discrimination by infrastructural providers could offer important provisions for freedom of expression, particularly for groups that are at risk of facing systemic discrimination.228 And while recent content moderation battles have involved infrastructural services—such as Amazon Web Services removing Parler after the January 6 insurrection229—the risks posed by inappropriate moderation among infrastructural services are high and may generally outweigh the benefits gained by enabling continued discretion for the reasons outlined above. Both deeper empirical study and a broader normative conversation about these choices are necessary, but such a choice would favor equitable service provision over maintaining the ability for providers to discriminate, even against heinous but lawful content.
Online infrastructure services were developed under the intermediary liability protections of Section 230 of the Communications Decency Act, which allows online services and internet users to provide and moderate content from others without being held liable for third-party content or their good faith moderation decisions.230 Recently, Section 230 has been targeted for reform by liberals seeking to combat disinformation on social media platforms and by conservatives for alleged censorship or suppression of conservative voices on those same platforms.231 Some of these proposals have been well-tailored to particular areas of concern, while others have been disconnected from the actual changes they might precipitate. In either case, significantly changing Section 230 protections in response to problems in the consumer-facing application layer may result in significant negative disruptions to many online services and likely have unintended consequences for infrastructural services. Reform or repeal of Section 230 would powerfully affect the speech of individual users—particularly those from traditionally marginalized groups—not just on gatekeeper platforms, but also across every website, app, and backend service. CDNs, cloud hosts, and other infrastructure providers may face the reality that their business models are no longer viable due to liability issues. Few proposing reform or elimination of Section 230 wish to unintentionally incapacitate such a wide array of everyday services, which is why recent proposals are increasingly exempting infrastructural players.232
Thus, the federal government must simultaneously enhance areas of responsibility for online infrastructure services and implement protections to ensure that these services are not unintentionally upended by any potential intermediary liability changes. Taking lessons from the benefits and challenges that have emerged over the years around Section 230, the public interest requires a mix of protections and obligations that are appropriate specifically for infrastructural services.
Online infrastructure has changed significantly over the past 20 years and continues to evolve. Today, the current crop of cloud technologies allows for numerous backend infrastructure tools—such as remote storage, raw computing, and proprietary software accessed remotely—to be offered as on-demand services that power everything from small-business websites to the largest online platforms. Although recent trends have been toward concentration and centralization of the cloud, the push for data localization and rise of 5G connectivity, edge computing, and blockchain suggest the possibility of more decentralized infrastructural services. Regardless, the demand for computing power continues to grow, and expectations of instantaneous and global access will drive continued evolution in infrastructure technologies.
In light of these crosswinds, drawing a precise, static line in the stack for which online services should be considered infrastructure is challenging233 and may only grow more difficult. To ensure that this model encompasses the companies and products whose goal is to operate as infrastructure, without unnecessarily impeding innovation or market evolution, this proposal would allow businesses to self-select and opt in a business line to the new regulatory model, subject to meeting designated structural requirements and approval by the regulating entity.
An opt-in approach offers a degree of future-proofing that may be difficult to provide through statutory definition alone. Allowing companies to generally opt in ensures that only those that consider themselves infrastructural and understand the requirements choose this model; this is expected to comprise a minority of online services overall. This approach may enable new companies to start with the explicit goal of competing as online infrastructure and would offer all infrastructural companies a strong defense against the technical, legal, and public relations costs resulting from good and bad faith demands for increased content moderation lower in the stack. While challenges exist around the incentives and consequences for infrastructure providers in and outside of the tier, those opting in would be regulated by an entity that prioritizes the goals of online infrastructure. Infrastructural firms outside the tier will have to deal with rules designed for broader online services or gatekeepers and deal with the business realities of any potential intermediary liability changes. Business customers will be able to exercise choice in determining which online service provider may best meet their infrastructural needs.
Even with an opt-in approach, regulators will need criteria for eligibility to help clearly describe the target audience and guide them in preventing abuse. Given the variety of service architectures in this space, flexible criteria are key. Administrators of the online infrastructure rules, in conjunction with technology experts at the National Institute of Standards and Technology (NIST), should be tasked with helping to assess qualifying online infrastructure services and developing such criteria. Some, but not necessarily all, of those characteristics of online infrastructure could include:
This report does not envision that payment processors or decentralized payment platforms would be eligible to opt in to the online infrastructure class in its initial establishment. Financial transactions are too enmeshed in the existing financial regulatory system, fraught with broader policy implications that require deeper consideration outside the scope of this report.
The business lines that opt in to the online infrastructure tier would not be subject to the general online services and gatekeeper regulations described later in this report, although select rules may be mirrored from those tiers. Notably, there is a strong incentive for companies to opt covered business lines in to the dedicated online infrastructure regime. If they do not do so, and have been designated as a gatekeeper, an infrastructural business line would be fully subject to gatekeeper regulation in addition to general online services regulation; alternatively, that business line, if sufficiently dominant, could qualify the entire company as a gatekeeper.
Companies with existing online infrastructure business lines, including those run by entities that will be designated as gatekeepers, will need to undergo a specific process to bring that business line into the opt-in online infrastructure model. First, companies would need to engage in a series of steps to separate the data and operations of the designated online infrastructure product from their other lines of business. In some cases, the easiest way to do this may be to establish an entirely new, separate entity. Such an approach is not without precedent. For example, the National Bank Act allows banks only to engage in certain prescribed activities; if a bank wants to engage in nonbanking activities, it has to create a holding company and create affiliates to do the nonbanking work. Importantly, any interaction between the bank and nonbank entity would need to be an arms-length transaction, as if they were doing business with an unaffiliated firm. Similar restrictions could apply here.
Second, strict data confidentiality restrictions for companies with business lines in multiple tiers will also be put in place. The entity would have to agree to these restrictions by attesting to them at the corporate and CEO level, with substantial civil and criminal penalties for the company and executives. Third, the regulators responsible for overseeing online infrastructure can veto an entity that they believe has not met the requirements and potentially even conduct a renewal process periodically to ensure appropriate inclusion. Finally, if an entity that has previously opted in to the online infrastructure class wished to exit that tier, there would be a required exit period of at least three years, and the company would be regulated as a general online service upon exit and assessed for potential gatekeeper qualification.
While business lines in the opt-in online infrastructure tier would not be eligible for gatekeeper status, in the tradition of infrastructural regulation, there may be individual services or submarkets of online infrastructure that become so critical to the operation of online infrastructure that they mirror traditional essential or critical infrastructure. Therefore, an essential or critical infrastructure designation for certain services or submarkets should be explored in conjunction with the U.S. Department of Homeland Security’s Cybersecurity and Infrastructure Administration and subject to additional requirements if so designated, with the specific goal of ensuring reliability, security, and access. Related strategies are already being explored around cloud systems in the financial services sector, including the request of Reps. Katie Porter (D-CA) and Nydia Velázquez (D-NY) to the Financial Stability Oversight Council to designate cloud storage providers as systemically important financial market utilities.234
Inspired by the robust progressive tradition of protecting the public interest in critical parts of the economy, the foundational responsibilities imposed under this tier are common carriage and interoperability requirements to ensure fair treatment of competitors and adjacent services. This report focuses on two components of common carriage: access and nondiscrimination. Specifically for online infrastructure services such as web hosts, this will require both access to the service by any lawful entity for lawful content and nondiscriminatory pricing of the services; that is, different classes of hosting services may have different prices, but hosts cannot price discriminate among customers within those services. Hence, incorporation of common carriage requirements protects against both content and economic discrimination.
While the tenets of mandated common carriage and content nondiscrimination do not make sense for every online service, they are appropriate for online infrastructure. Whereas there are times where higher-stack, consumer-facing online services need to exercise greater discretion and assume greater responsibility for the activity enabled by their online business products, lower-stack infrastructural services generally need the opposite: protection from any increasing liability for the consumer-facing services they enable and requirements to deal equitably among legal content and customers.
For online infrastructural services, common carriage and greater intermediary liability protections should go hand in hand. This would eliminate the ability for infrastructure providers to arbitrarily discriminate among their customers, compel them to carry all legal content, and insulate them from inappropriate legal liability or evolving changes in liability for their service provision. Legal content does not mean all content. Content that violates federal law such as child sexual abusive material,235 accounts from designated foreign terrorist organizations,236 or content in violation of sanctions programs or from sanctioned countries237 would continue to be prohibited and require legally appropriate preventative measures. In extremely narrow cases, Congress could allow the regulator to identify other specific exemptions.
The tradition of common carriage in the telecommunications space also provides useful references for conceptualizing appropriate public safety obligations. Telecommunication providers are required to facilitate emergency services such as 911 connections and emergency information. In a similar vein, services opting in to the online infrastructure model could be required to provide certain standardized public safety measures. These could include clear requirements and standards on reporting illegal content, such as incitement to violence, time and process requirements for the service to investigate and act, referral requirements to appropriate law enforcement entities, and standard lawful preservation requests.
The goal of the online infrastructure class is to ensure the continued fulfillment and delivery of online services. This does not just mean technical rules around ensuring stability or security but ensuring that customers have broad ability to prevent legal, pricing, or technical lock-in. For example, an online infrastructure regulator could not only ensure clear price transparency and billing recourse but also prevent abusive contractual, technical, or economic lock-in costs.
Beyond the common carriage obligations addressed below, additional obligations imposed under online infrastructure’s public interest bargain might include cybersecurity standards, privacy rules, and environmental standards, particularly given the significant and growing environmental impacts of data centers that are too often overlooked.238
For regulators, being able to craft more stable rules specific to strengthening and protecting online infrastructure should be easier than crafting broader rules for online services, and indeed should remove a roadblock to crafting appropriate regulations for consumer-facing services. Under the status quo, much of the legislation targeting online services—which is primarily conceived for consumer-facing rather than infrastructural services—would need to include infrastructure exemptions to target restrictions to have their intended effects. In contrast, the Center for American Progress proposes distinct tiers that disentangle fundamentally different roles in online services provision and should enable better regulation with fewer unintended consequences in both cases.
The speed, scale, and novelty of some online services markets have at times challenged regulators, but so has the deliberate obfuscation and misdirection by online service providers seeking to evade government oversight, sometimes relying on exaggerated claims of their own complexity or novelty to defend against regulation. Industry speed and complexity, however, are not insurmountable challenges to restoring democratic oversight. The federal government needs to establish sufficient expertise to assess, communicate, and regulate, on behalf of the American public, the practices, technologies, benefits, and harms associated with online services. To anticipate and encourage innovation while protecting the public interest, significant regulatory enhancements and statutory protections are needed to protect consumers, safeguard civil rights, and promote competition online.
For general online services—all services outside of the online infrastructure tier proposed above—this section proposes to enhance the following tools and capabilities:
In enforcing these rules, robust options should be put on the table: civil enforcement, criminal enforcement, fines, and criminal penalties for executives will all have a role to play in enforcing the general online services rules promulgated under the new model.
A mandate for ongoing oversight, new investigatory responsibilities, and enhanced disclosure and transparency powers are an essential foundation for online services regulators. At present, the stark asymmetry in knowledge and data about online harms between online services companies and everyone else means that regulators, academics, journalists, and the public are at a significant disadvantage in trying to understand even the broadest strokes of issues online. The ability to investigate, audit, and publish reports about online services markets and emerging issues, such as bias in machine-learning programs or the impact of algorithmic amplification on disinformation, would allow for significantly improved transparency and understanding. Specific obligations should be given to regulators to investigate emerging problems and problems for which online services regulators may be the sole oversight body.
Research and investigatory authorities would need to be sufficiently broad to encompass products, features, algorithms, business practices, or other technical architectures of online services. Specialist regulators could produce new reports or impact assessments on these issues, which should be public by default and could serve as a starting point for sector-specific discussions, codes of conduct, or traditional rule-making processes. These could benefit the wider ecosystem of concerned parties—including Congress, peer agencies, and the courts—in seeking to understand and remedy consumer and competitive harms over time.
Newly empowered regulators can also play a role in shepherding data disclosure frameworks and standards that support researchers, global regulators, and the public in pursuing independent research of online services markets. Increasing access to academics and other qualified researchers is one of the few ways to allow for greater understanding of online services. There has been recent momentum in the EU and the United States to better address this information disparity, including the EU’s Access to Data Held by Digital Platforms for the Purposes of Social Scientific Research working group,239 new proposals for research of very large platforms in the EU’s Digital Services Act,240 and the United States’ proposed Social Media Disclosure and Transparency of Advertisements Act.241 Similar efforts to allow research access should be incorporated as part of general online services oversight responsibilities. Further coordination with global regulators around data disclosure standards could be particularly productive in catalyzing global understanding of shared challenges. Across research and investigation areas, regulators could develop and abide by standards and best practices for appropriately navigating the substantial privacy and intellectual properties sensitivities at play.
In exercising new and enhanced investigative, oversight, and data disclosure authorities within online services markets, regulators may encounter issues that are pertinent to other state or federal agencies. To maximize both new and existing agency powers in service of the public interest, regulators should have referral powers to bring issues, conduct, or other pertinent information to the attention of other government agencies. Referrals should always be made wherever potentially unlawful conduct is found under other existing statutes.
In many cases, sector-specific regulators may currently be responsible for regulating activities occurring in part on online platforms—such as advertising employment opportunities—without the informational awareness or technical expertise to fully address them. For this reason, the entity charged with administering the general online services regulatory model should serve as both a center for expertise on online services and a partner to other agencies as they include aspects of online services in their sector-specific regulations. In so doing, the general online services regulations can provide common elements and principles for integration. These collaborations could deconflict online services regulatory approaches across the whole of government. As noted above, online services regulators would also have robust investigative and referral powers and could shine a light on problems for other agencies to address.
For example, existing sector-specific regulators, such as the National Highway Traffic Safety Administration (NHTSA), oversee self-driving cars and automated vehicle technologies.242 As these efforts evolve, NHTSA could leverage online services principles, standards, or regulations in their efforts, or at minimum, ensure that they are not in conflict. Similarly, an online services regulator that discovers issues in the online services used by self-driving cars, such as insecure software or deceptive claims of system performance, would be able to refer them to NHTSA and, if needed, coordinate on addressing identified harms.
Through these referral and collaboration processes, regulators can assist a range of federal and state bodies in making existing regulations robust against emerging and future problems. There is ample precedent for such an approach. For example, the Federal Deposit Insurance Corporation is responsible for supervising all chartered banks, with a range of accompanying enforcement powers to do so. It examines these banks not only for safety and soundness, but also for compliance with more than 20 different federal consumer protection statutes—even statutes for which other agencies are the primary regulators. In areas where no other entity has responsibility to guard the public interest, dedicated rule-making responsibilities will enable online services regulators to ensure that challenges raised by online services do not fall through the cracks.
For online service providers that do not choose to be classified as online infrastructure, Congress should empower an expert administrative body with robust oversight and rule-making powers guided by the legislatively enumerated regulatory principles described below. (A discussion on selecting a body continues below in “Administering regulation.”) However, as general principles on their own are insufficient to guard against industry capture and ensure clear administrability, each category should also include more specific per se violations that create clear guardrails. For example, the House Judiciary Antitrust Subcommittee’s tech antitrust bills and new Senate companion bills243 put forth a number of rules around nondiscrimination that would make this behavior unlawful for dominant online platforms; such a rule exemplifies the type of guardrail where Congress fully understands a problematic activity at the time of drafting a statute. Further work will expand on proposed clear-cut statutory rules and examples of rule-making additions for each of these spaces.
Critically, this regulatory model would have the ability to create rules for all players in online services markets. Much of the focus on curbing the tech sector’s abuses have centered on the largest gatekeeper companies. However, abuses can happen from all players in the ecosystem, including users, vendors, advertisers, and smaller business. Where harms are widespread, having rules that apply to everyone is critical to ensuring the development of a stronger, safer, and more vibrant online services ecosystem.
Thus, some new statutes and accompanying rule-making activity may create clear rules and regulations for online services generally. But much of this activity would likely be targeted at specific markets, technologies, or submarkets. Within the broader ecosystem of online services, it is likely that different types of online services will require different sets of rules for their individual, sector-specific markets. Problems in the augmented reality market, for example, will likely require some different rules and remedies than markets for IoT devices or AI. As a starting place for oversight and rule-making, Congress should designate initial sector-specific markets where it suspects regulatory intervention is overdue based on the principles below. The entity or entities in charge of administering this regulation should be able to select future markets for this purpose. Specific selection criteria for initial focus markets will be the subject of further work. 
The Center for American Progress’ proposed approach is a hybrid one: Congress would both define specific practices that would be explicitly outlawed and enumerate broader principles around which regulators could interpret and craft rules. For example, Congress might explicitly prohibit the sale of biometric data but also more generally prohibit insecure and data extractive practices; a regulator might then promulgate a rule prohibiting firms from scraping photos from online services without securing users’ permission to build a facial recognition service for clients. The combination of clear guardrails and the flexibility of a principles-based approach offers flexibility to address future problems and mitigation of any industry capture of the regulator.
Similar principles-based approaches have been used to regulate financial services. Building on a robust separation regime, the federal government empowers financial regulators with flexible rule-making capabilities, in theory enabling them to address unanticipated, emergent threats. Agencies charged with regulating financial institutions or their products—such as the Federal Reserve and the Consumer Financial Protection Bureau—have been empowered with broad supervisory and regulatory authority to ensure safety and soundness and guard against unfair and deceptive practices, respectively. This is critical because of the complexity of the financial sector and the inability for Congress to enumerate all of the potential mechanisms that could lead to systemic risk or consumer harm. Likewise, the speed of technological development in some online services markets can compound the existing challenges of needing specificity and specialized expertise, making multiple, flexible tools and ongoing oversight and rule-making responsibilities essential.
An online services regulator would promulgate rules in varied digital markets where harms arise on the basis of broad, statutorily-defined prohibited practices and factors that must be considered. Congress would describe these general categories of prohibited behavior—in addition to specific practices defined in statute to be unlawful for online services—and regulators would continue their work by developing and applying rules to specific technologies, practices, or markets, appropriately considering the requisite factors named as process requirements as new issues arise. Some of these principles intentionally overlap existing statutory areas. As noted under referral powers, many of these areas have yet to be clearly applied to the online services space, and new regulators will be a force multiplier in filling regulatory gaps and updating existing statutes. The first set of principles would take the form of unlawful prohibitions:
Furthermore, the list of principles could include some combination of the following factors that regulators are required to consider in the development of specific rules. In many cases, rule-making can simultaneously advance all of these principles, but in others, explicit balancing across principles and weighing of any tensions or trade-offs will be required. In addition to the lawful prohibitions outlined above, this proposal aims to outline the start of an affirmative public interest mandate for online services regulation. Including such public interest considerations furthers the robust regulatory tradition of explicit, affirmative public interest obligations for U.S. regulatory bodies:
Dedicated, expert regulators would promulgate rules that consider these factors when seeking to operationalize the classes of prohibitions enumerated by Congress, turning general principles into rules that address the specifics of different technologies and online services markets. Regulators will invest in expertise to understand the complexity and variety of the online services space, bringing together the technical, social, economic, and legal knowledge required to effectively govern it. In doing so, they will be informed and engaged enough to strike an appropriate regulatory balance among risk, growth, and competing objectives in burgeoning online services industries. They will identify common problems across disparate industries and consult with other regulatory bodies in their own rule-making. Together, this should give regulators sufficient rule-making authority over online services and the expertise necessary to govern them effectively. These authorities would complement, not supplement, existing FTC, DOJ, and other sector-specific jurisdictions over related issues.
As noted, while some rules may apply to all general online services, many will likely be targeted by a specific technology or submarket. Looking ahead, it is easy to imagine instances where principled rule-making and expert oversight could be used as appropriate complements to address the economic, privacy, consumer protection, and civil rights harms outlined earlier in this report. To help bring to life new rule-making powers, consider the below examples:
A few companies achieved early success in new sectors of the digital economy and now dominate much of the consumer internet. Recent scholarship on digital gatekeepers has disentangled the natural features of digital markets that support capture and tipping from the anti-competitive actions companies may take to exploit those features. Digital markets are characterized by strong network effects, extreme economies of scale and scope, high barriers to entry, and acute information asymmetries between dominant platforms, dependent players, and Americans who rely on these services every day.245 While the individual characteristics of digital markets are nothing new, the combination of these features and the degree to which they play out online poses acute challenges to market function and heavily favors dominant digital incumbents.246 In the vacuum of regulation and antitrust enforcement in recent decades, some dominant companies have abused natural market conditions to kill competitors, entrench dominance, stave off regulation, chill innovation, and continually leverage their dominance into adjacent sectors of the economy. Consumers—and the hundreds of thousands of small and medium businesses and digital creators that are reliant on these companies to reach them—have lost out.
Looking solely at economic outcomes does not fully capture the risks posed by digital gatekeepers. Very large digital gatekeepers are systemically important, as their actions have major implications for the U.S. economy, society, and security. Similar to systemically important financial institutions, they pose widespread risks given their status as functionally essential and ubiquitous informational infrastructure. Abusive behavior toward consumers, potential competitors, or corporate negligence on a range of critical public interest issues—such as cybersecurity, data privacy, discrimination, political advertising, content moderation, and site reliability—can generate cascading social harms and significant economic costs. Americans want, and regulators should provide, oversight over digital gatekeepers that have the potential to cause massive harm—especially in those areas where harms are unseen or diffuse.247 Given their scale, such platforms should not be allowed to operate in ways that are fundamentally contrary to the public interest. While a robust and complex conversation is needed on what should constitute appropriate and rights-respecting forms of oversight and regulation in response to risk, the need for oversight is clear. The EU’s Digital Services Act,248 a recent legislative proposal that explicitly incorporates language around risk and due diligence for very large online platforms, grapples with similar issues.
The Center for American Progress has previously written in favor of more aggressive antitrust action249 and robust competition policies,250 including statements of support for the package of antitrust and competition bills reported out of the House Judiciary Committee in the 117th Congress that tackle digital gatekeepers. These bills, led by Chairman David Cicilline (D-RI) and informed by the subcommittee’s report on digital platforms,251 are a direct response to the problems outlined above. The Ending Platform Monopolies Act, which addresses underlying conflicts of interest between platforms and commerce, provides a unique opportunity to change foundational incentives for dominant platforms, rather than needing to monitor behavior in-depth on an ongoing basis. The American Choice and Innovation Online Act and new companion legislation led by Sen. Amy Klobuchar (D-MN) would introduce nondiscrimination provisions and restrictions on self-preferencing, offering an opportunity to curb many of the competitive abuses to small businesses outlined above.252 These bills present the 117th Congress with a meaningful opportunity to address some of the most troubling abuses of gatekeeper power.
Considering the importance of gatekeeping firms to the U.S. economy and the propensity for digital markets to tend toward tipping, the U.S. should consider additional laws and regulatory scrutiny of such firms going forward. Indeed, because antitrust enforcement in digital markets has been so anemic, it is possible that even after separation, a number of firms would still qualify as digital gatekeepers as conceptualized below. Additional regulatory scrutiny can ensure a range of harms and risks are sustainably addressed, including and beyond those firms that have gatekeeping power but could be dealt with through structural approaches. To determine which firms deserve additional regulatory scrutiny, this report builds on existing scholarship—including the House and Senate tech antitrust package and the EU’s Digital Markets Act—to propose a test that could be used to identify future gatekeepers. Following the test, this report suggests further regulation and risk mitigation tools that may be brought to bear on qualifying gatekeeper services. 
Recent scholarship and government investigations from around the globe have coalesced around a set of factors that characterize digital gatekeepers, but articulating a precise definition is the subject of ongoing legal and policy work. As described by FTC Chair Lina Khan, “Gatekeeper power can arise any time there is a network monopoly, a feature of industries with high fixed costs and network effects, or the phenomenon whereby a product or service becomes more valuable the more that users use it.”253 In the context of online services or tech platforms, she notes, “While more extensive studies of platform power would benefit from being platform-specific, identifying the common bases of their dominance helps place them within existing legal frameworks. The first is gatekeeper power. This power stems from the fact that these companies serve effectively as infrastructure for digital markets—they are distribution channels, the arteries of commerce. They have captured control over technologies that other firms rely on to do business in the online economy.”254 Acknowledging the divergent forms of gatekeeping that arise from a similar set of conditions in digital markets, a test that covers multiple conditions is preferable to a single cookie-cutter definition.
In order to formulate such a test, this report surveyed major research reports and legislation targeting digital gatekeepers in recent years: the University of Chicago Stigler Center Committee on Digital Platforms report,255 U.S. House Antitrust Subcommittee report,256 U.K. Digital Competition Expert Panel report,257 U.K. Competition and Markets Authority report,258 Australian Competition and Consumer Commission digital platforms inquiry report,259 French Competition Authority digital platforms report,260 Germany’s 19a antitrust policies,261 European Commission’s digital era competition report,262 EU’s Digital Markets Act,263 Japan’s Ministry of Economy, Trade, and Industry digital platforms regulation,264 Harvard Shorenstein Center proposal,265 Brookings Institution big tech regulation report,266 Public Knowledge Digital Platform Act,267 and Washington Equitable Growth antitrust and competition report.268 This report catalogued the definitional proposals across reports and identified the named characteristics for harmful digital gatekeeping in each.
Building on this foundation, each characteristic was considered in light of their regulatory feasibility in the United States. After exploring various approaches to combining these factors—comparing them with the existing landscape of digital gatekeepers and imagining the likely gatekeepers of tomorrow—the authors developed the following digital gatekeeper test to identify gatekeepers that pose unacceptable risk to the public interest. The EU’s Digital Markets Act and the U.S. House tech antitrust legislative suite,269 released midway through this report’s own development, provided useful comparisons, and some elements are echoed in this proposal.270
In order to be subject to gatekeeper regulations, firms would first have to meet the definition of online services in a primary business unit.271 Many qualifying gatekeepers will operate multisided markets, but other large, dominant online businesses or ubiquitous, systemically important services may also qualify. Eligible companies will qualify for gatekeeper status if three of the four conditions below are met. While each is an important part of the puzzle, every condition also has a predictable counterexample that means it is not necessary to establish gatekeeper power in all cases. Due to the extreme economies of scale and scope in digital markets and the prevalence of cross-subsidized business lines, either the provider overall or only one of a provider’s business lines needs to meet the requirements. These thresholds are provided for illustrative purposes; whatever criteria are chosen, regulators will need the ability to update specific quantitative thresholds based on market developments and economic changes over time. These ideas are meant to be further data points in the continuing conversation about how best to turn well-established economic observations and criticisms on digital gatekeepers into a functional, empirical regulatory test that effectively targets products with harmful gatekeeping power.
Table 1
Condition #1 accounts for an online services provider’s sheer economic importance to the United States. The first threshold of $9 billion in U.S. annual revenue in the past three years and the second threshold of $90 billion in fair market value or average U.S. market capitalization in the past financial year mirror the EU’s Digital Markets Act proposals around economic importance, scaled to U.S. market size; the Digital Markets Act threshold of €65 billion used the average market cap of the EURO STOXX 50, adjusted for future growth expectations. The House’s American Choice and Innovation Online Act, which set the threshold at $600 billion, does not present a direct comparison here given its global scope. Such a figure might be indexed to growth or calibrated to consistently capture above-average market caps for a group of the largest U.S. companies over time. In considering potential counterexamples, a product’s technological importance might grow more quickly than its economic performance, and thus monetary impact on the national economy may not always accurately reflect a product’s true importance in the short run.
Condition #2 accounts for products whose market power affords them gatekeeping power. Firms can qualify either with a market share above 30 percent or a Q ratio, also known as Tobin’s Q, that is consistently higher than 2. At 30 percent market share, this threshold aims to capture not only firms that hold monopoly or duopoly power, but also firms that dominate markets as part of an oligopoly. However, market share has increasingly come under scrutiny as a standalone indicator of market power. And online services firms have tended to complicate the process of market definition through multisidedness, zero-price services, and data economies. The variation in business models among digital gatekeepers further indicates that multiple indicia of market power may be needed. Thus, a firm can also qualify by exhibiting a Q ratio greater than 2.
Tobin’s Q is the firm’s market value divided by the replacement cost of the firm’s capital assets.272 As explained by Marc Jarsulic and others, persistently high Q ratios are one indicator that a firm is extracting monopoly rents.273 When financial market valuations exceed capital costs, there is an incentive to buy assets and employ them in that line of business. In a competitive market, entry should continue until increased supply lowers returns and the Q ratio declines to one. The Peters-Taylor methodology of calculating Q, which includes intangible capital in the denominator of Q, is a sensible approach to valuing online services companies, for which intangible assets are significant.274 Setting the threshold at 2 essentially says that if a company is persistently valued by financial markets at a level that is more than double the worth of its assets, firms are earning economic rents that indicate significant market power.275 Given the free and multisided nature of many digital services, Q ratios are a useful metric for understanding the market power of digital firms.
In addition to these two indicators, regulators should be empowered to establish more stringent standards through rule-making processes as markets evolve. The Digital Markets Act and the American Choice and Innovation Online Act both acknowledge the importance of market power generally. Most gatekeepers will likely have significant market power, but there are cases where gatekeeping power may arise prior to a product’s market capture or absent a clearly definable market; therefore, condition #2 is expected in most, but not necessarily all cases.
Condition #3 speaks directly to a product’s intermediation power. The threshold seeks to target services that play economically important intermediation roles based on high numbers of monthly active users and business users or based on holding significant market power in a strategically critical market. This proposal’s threshold for the number of U.S. business users—10,000—reflects the EU’s business threshold in the Digital Markets Act, which was itself calibrated to represent a small share of the entire population of “heavy” business users on EU platforms. The number of monthly active U.S. users, 30 million, represents approximately 10 percent of the U.S. population, which is also slightly below the threshold in the American Choice and Innovation Online Act of 50 million monthly active U.S. users. The threshold for critical buyer or seller market share, 50 percent, represents a point above which courts have tended to find monopoly power.276 Further research is needed to identify the particular characteristics of economically powerful and systemically important digital intermediation services, but in the absence of clearer tipping points, a descriptive but conservative approach could be a good starting point. A small but ubiquitous and highly interconnected firm may exercise robust intermediation power, and whatever metrics are chosen should reflect this possibility. Most digital gatekeepers are expected to meet condition #3, but it is possible that an online service could acquire gatekeeper influence through sheer economic size, market dominance, and market durability, rather than having a literal intermediation position.
Condition #4 introduces the traditional “durability” metric, scaled to three years to account for the swift pace of digital markets, as a final indicator of gatekeeping power. The Digital Markets Act proposes a similar approach to durability. Incredibly popular products could foreseeably arise and amass gatekeeper power in less than three years. Thus, condition #4 will likely be met in many but not all cases.
Figure 2
In crafting regulations, Congress and relevant regulators—whether through the pending tech antitrust bills or as part of a future comprehensive regulatory statute—should reign in anti-competitive practices and consumer harms that are uniquely enabled by gatekeeper status, above and beyond any rules proposed through general online services regulations. Specifically, goals for gatekeeper regulation should include:
To achieve these gatekeeper goals, both structural and functional separation should be utilized where appropriate. In her 2019 article, “The Separation of Platforms and Commerce,” Khan described operational or functional separation and structural separation as such:
An operational or functional separation requires the firm to create separate divisions within the firm, requiring that a platform wishing to engage in commerce may do so only through a separate and independent affiliate, which the platform may not favor in any manner. A full structural separation, by contrast, requires that the platform activity and commercial activity be undertaken through separate corporations with distinct ownership and management.277
Khan makes a persuasive case for reviving structural separation approaches for digital markets, particularly in light of the challenges in functional separation.278 Indeed, when determined by the courts or clearly laid out in statue, structural separation is a preferred strategy for grappling with fundamental anti-competitive conflicts of interests within digital platforms, although it has been infrequently used in recent decades.
Structural separation approaches for gatekeepers should continue to be initiated through antitrust litigation or in clearly defined statutes, such as in the proposed Ending Platform Monopolies Act.279 A regulatory entity with the ability to determine structural separation without clear statutory guidelines, however, raises the risk of potential abuse and inevitable legal challenges.280 Rather, additional tools for an online services regulator may include reviewing or overseeing, but not initiating, separation regimes.
Agency regulators tasked with gatekeeper oversight might also be given the ability to initiate functional separation. In some circumstances, functional separation could be the maximal appropriate tool for a dedicated regulator to level on a gatekeeper—preserving structural separation as an option through antitrust enforcement or when explicitly directed by statute in cases of unavoidable conflicts of interest. Functional separation, with its tradeoffs, may be less effective and more difficult to administer than structural separation, but should be preserved as a tool in the regulator’s toolbox when structural separation may not be possible.
Beyond structural and functional separation approaches, further strategies employed in gatekeeper oversight should include restrictions on self-preferencing, bundling, price discrimination, interoperability and data-sharing rules, and enhanced disclosure obligations tailored to particular forms of gatekeeping power, among others:
While some of the harms from dominant digital gatekeepers can be alleviated by industrywide rules and baselines imposed on general online service providers, others will require dedicated scrutiny. Due to the regulatory debt built up around gatekeepers and their economic importance, one-size-fits-all rules cannot always capture the issues posed by particular gatekeeping powers. In addition to clear rules wherever possible, a regulator with the mandate to examine the best way to preserve and balance the principles set forth by Congress, with the flexibility to evaluate specific types of gatekeeper power, would be a useful complement to reinvigorated antitrust action in the digital platform space.
The instantaneous speed, amplification, discovery, and relational nature of speech online are at the heart of the productive and transformative capabilities of the internet. But the rise of gatekeeper platforms means that many of the world’s online interactions are intermediated by a handful of companies whose decisions profoundly affect how people communicate and whose business incentives cut against the public interest. In particular, negligent business models that amplify, promote, and target content to certain users are accelerating societal divisions, compounding existing inequities, and sustaining extractive surveillance business models. These systems have been easily exploited by malicious actors for purposes of harassment, voter suppression, and disinformation, adding even greater urgency to long-standing problems.
The regulatory proposals in this report and the relevant tools outlined below focus on how online services treat consumers, creators, business users, workers, and competitors. They do not constitute a program of direct speech regulation. Instead, this framework creates the capacity for regulators to identify clear risks, including systemic risks, and address them through the lenses of consumer protection, civil rights, and competition. These are long-standing regulatory and oversight traditions that can address specific issues as they are—for example, abusive business practices, deprivation of rights, or anti-competitive practices. They offer sensible, legal interventions related to many, though certainly not all, issues included in the locus of discussion around so-called harmful but legal online content. These traditions may offer a more tractable lens than that of speech regulation, which will face steep challenges in the courts.
Many of the tools below target the troubling information asymmetry between private online services providers and everyone else. These providers are profoundly influential in shaping public discourse. Their refusal to provide the public, regulators, or researchers with even basic data about their business practices and programs is a foundational rejection of their public interest responsibilities. The apparent misrepresentations that some providers have willingly made to the public and even Congress are likewise deeply troubling.282 A regulatory entity with the tools to bring transparency, research, and understanding to this space can be catalytic in illuminating potential remedies.
In unpacking how those tools apply to harmful content online, it is helpful to disaggregate the term into specific issues. Myriad problems tend to get lumped together in this discussion, particularly as they relate to reform or repeal of Section 230 intermediary liability protections. These include but are not limited to harassment, hate speech, scams, discrimination, fraud in advertising, doxxing, algorithmic transparency, misinformation, disinformation, voter suppression, radicalization, election interference, or nonconsensual pornography. Each of these is a serious issue meriting dedicated consideration. Americans’ online interactions are real, innumerable, and complex: Legal and regulatory systems do not try to address embodied actions of harassment, voter intimidation, discrimination, and public health interference with a single policy response; neither should they do so with their digital instantiations. To better craft policy responses to these issues, regulators must widen their aperture beyond a flattened idea of “content” to look upstream at business models and design choices, downstream at the relational harms or other direct impacts on individuals, and broadly at the historical context and information environment that online service platforms create.283 Section 230 is one part of this ecosystem—and perhaps the most discussed given that Congress does have leeway under the First Amendment to change intermediary liability rules—but is not the whole ballgame.
New regulatory tools might be able to address various aspects of specific problems, such as deceptive design, negligent business practices, infringement of rights, abuses of market power, or the need for greater redress or understanding. An online infrastructure regulator might impose public interest transparency reporting or staffing obligations to enable rights-respecting treatment of illegal content among infrastructural providers. A general online services regulator might bring to bear enhanced investigatory and rule-making powers wherever technologies or business practices are anti-competitive, unfair, deceptive, abusive, insecure, data extractive, or likely to violate civil rights. The requirement to consider effects on information diversity within online services rule-making will ensure that new rules promote pluralism over continued concentration. Additional gatekeeper rules for the largest and most important players will provide further oversight and systemic mitigation of risks to national cybersecurity and democratic infrastructures, among other critical systems.
For purposes of illustration, concrete uses of these tools might include:
Expertise, investigations, and referrals: As part of their oversight obligations, regulators may conduct investigations into critical issues and, where necessary, serve as an expert partner on efforts by other government entities in understanding and protecting the public interest in their areas of work. These inquiries may touch on various issues related to harmful online content, such as understanding online voter suppression or foreign influence operations, in partnership with relevant agencies such as the DOJ or FEC.
Going forward, First Amendment protections, the risk of government abuse in speech regulation, American values, and the history of the government’s failure to protect the expression of oppressed Americans should inform a difficult calculus for online speech regulation. While the goal of protecting freedom of expression is clear, the best means to achieve it is an open debate. How should society treat speech that seeks to undermine the very idea of public discourse itself? What is an appropriate, rights-respecting role for government within that treatment? A favorite U.S. adage suggests that the best antidote for harmful speech is more speech. But automated, instantaneous global amplification and surveillance-driven targeting that are used to uplift, silence, or drown out other voices begs the question of whether protecting freedom of expression requires approaches that accommodate, rather than ignore, the ways technology has changed how people communicate.
An entity with the tools to bring transparency, research, and understanding to this space is one that can illuminate potential remedies going forward. Especially as the conversation around reform or repeal of Section 230’s intermediary liability protections continues, gaining greater understanding of the impacts of reform can aid in aligning outcomes with policy goals. In particular, increased transparency on platform moderation practices will be needed to assess the effects of any carefully calibrated updates. Additionally, if there are very narrow, specific issues that are so severe that they merit changes to speech regulation by Congress, such proposals would only be upheld by the courts with overwhelming evidence of harm. Thus, the full range of possible actions in this realm will require clearer evidence on harms and tradeoffs. Greater understanding can aid lawmakers and the public in assessing proposals on their merits.
Even as the national discussion about the government’s role in guarding freedom of expression online continues, more clear-cut regulatory approaches grounded in civil rights, consumer protection, competition, and dramatically stronger transparency standards are appropriate responses in the immediate term. The proposals in this report outline a variety of sensible new tools for regulators to mitigate harm from online content problems. This rights-respecting approach is not totalizing, but it is a powerful, legal, and tractable place to start.
Regulatory effectiveness faces a host of challenges, including regulatory capture, enforcement failures, difficulty for users, and a range of capacity and cultural constraints.286 These factors present a strong argument for tools that are self-administering where possible, including structural separation and clear statutory lines for highly problematic practices. But as discussed above, there are limits to the ability of statutes to fully address the range, variety, and dynamism of some online services markets. Principles-based rule-making powers can offer a powerful complement to clear statutes in addressing complex, emerging issues and balancing conflicting priorities. New and existing statutes and rule-making powers will all need to be brought to bear in combination, despite the particular shortcomings of each. Shedding new light on longstanding administrability challenges is outside the scope of this paper. But going forward, these challenges should not be underestimated, nor should they serve as a barrier to action.
Expansion of existing agencies and consideration of new agencies should both be on the table. In either case, these proposals require significant expansion of the U.S. government’s capacity and expertise. Given the complexity of some online services—many of which deal in technical fields relating to software engineering, machine learning, or algorithmic design—and their direct impact on Americans’ access to opportunity, specialist regulators with appropriate sociotechnical expertise are required. The federal government must design a creative system that recruits needed expertise while sufficiently insulating agencies from industry capture. Such capacity will aid in making technologies more legible to the public, taking the air out of any unrealistic industry exaggerations of technical complexity and challenging unfounded objections to sensible regulation. Developing effective regulation will require wholesale rejection of the discriminatory industry dynamics—particularly around racial and gender-based discrimination—that are encoded and amplified throughout technologies, services, and products today.
Any additional responsibilities should be complementary and additive to existing DOJ, FTC, and FCC authorities, as well as sector-specific laws in other areas. Creating a center of excellence within the executive branch for online services could be a catalyst to ensure that the U.S. government can holistically, effectively, and consistently regulate new technologies. Specialist regulatory entities could also provide needed expertise and common principles for use in other areas, such as housing, labor, or transportation.
While some responsibilities described in this report mark clear shifts from current work at existing regulatory agencies, others are more natural outgrowths. Going forward, administrative options could include:
For illustrative purposes, a brief tour of options is below. 
Expand the FTC: An expanded FTC is in some ways a natural fit for these regulatory responsibilities. The FTC carries the dual mandate of competition and consumer protection. It looks holistically at these factors across large and small players and offers experience with issues around consumer data and privacy. And, as noted above, numerous proposals are already on the table to expand the FTC’s focus on data protection and digital markets, whether through its own rule-making or expanded powers in federal privacy or competition legislation.287 However, the agency is responsible for competition and consumer protection across many sectors of the economy. Especially in an era characterized by extreme corporate concentration across multiple sectors, the FTC is already vastly underfunded and understaffed relative to its mandate: As noted previously, over the past four decades, the U.S. economy has nearly tripled while FTC capacity was cut by more than a third.288 The FTC Office of Technology Research and Investigation has only a handful of staffers to support work across the commission.289 Adding a large new focus would necessitate a dramatic addition of resources and personnel. In recent history, the FTC also has been more of an enforcement agency than one that engages in rule-making, although it has some ability to do so in cases of demonstrated, prevalent problematic industry practices and where Congress has given it specific ability, as in the Children’s Online Privacy Protection Act.
Given the invisible yet pervasive nature of modern digital consumer protections harms and their threats to fundamental rights, the FTC will play a critical role in reigning in predatory practices, regardless of how any other expansions are accomplished.
Expand the FCC: The FCC’s roots as a telecommunications regulator tasked with common-carrier oversight suggest some relevance to administering the new online infrastructure model. The agency has significant rule-making expertise and staff technologists who understand the hardware and software sides of core communication technologies and lower-stack internet service providers. The agency may, however, be less well-suited to online services regulation more generally. Its work has historically tended to be deliberate, and it may face challenges in expanding to a broader role charged with competition policy, new technology markets, and dynamic regulation. Therefore, if distributing responsibilities, the more sensible option may be to split administration, housing online infrastructure oversight at the FCC and charging the FTC or a new agency with general online services and gatekeeper oversight.
Expand an existing executive branch agency such as the NTIA or NIST: The National Telecommunications and Information Administration (NTIA) is the executive branch agency charged with advising the president on telecommunications and information policy issues. While industry regulation of this scope and scale have traditionally been outside its mandate, particularly given the FCC’s authorities, the NTIA is among the agencies most familiar with internet governance challenges at home and abroad. Similarly, the National Institute of Standards and Technology (NIST) has begun to play a key role in setting cybersecurity standards, analyzing facial recognition, and beginning to outline the impacts of AI. Both the NTIA and NIST are part of the U.S. Department of Commerce, and while they have not historically held robust online services regulation roles, they are clear executive agency candidates for increased involvement. Housing new authorities at the executive agencies does, however, introduce greater risk around politicization and instability that may be precipitated by changes of administration. Thus, expanding the powers of an executive agency—rather than an independent one—would need to overcome steep administrability challenges and require strong congressional oversight.
Establish a new agency: Given the scale of distinctive expertise that effective online services regulation would require, a new agency may be a sensible path forward. A new body offers the chance to think carefully and creatively about administrative design without upending existing work. It enables a fresh start and dedicated focus, rather than adding a competing one; as Harold Feld notes in his writing on the question, expansion of existing agencies may pit the interests of the new focus against the old, where organizational culture and momentum strongly favor the latter.290 Other prominent experts and government officials studying the issue have increasingly determined that the challenges of digital markets may require a new entity.291 Tom Wheeler, the FCC chairman under President Barack Obama, along with Biden administration DOJ antitrust official Gene Kimmelman and former FCC official Phil Verveer, have proposed a new regulatory agency for digital gatekeepers.292 As a former chairman of the FCC, Wheeler’s recommendation for a new agency should be given some weight. Creating a new agency would demand significant resources and political will but may be the better long-term solution for the historic task at hand. 
Expand any future new privacy agency: There are several proposals before Congress to create a new federal data privacy agency, similar to the national data protection authorities found in most other countries. These proposals include Sen. Kirsten Gillibrand’s (D-NY) Data Protection Act293 and Reps. Anna Eshoo (D-CA) and Zoe Lofgren’s (D-CA) Online Privacy Act.294 If these bills were enacted and a new data privacy agency established, it may make sense to give the new body a hybrid mandate, to not only tackle privacy issues but also the interlocking economic, civil rights, and consumer protection concerns of the online services industry more broadly. Indeed, privacy is but one among several important areas of work around online services. Policymakers should take care not to solely prioritize privacy at the expense of other critical areas, such as competition, security, and expression.
All of the options outlined above have their advantages and disadvantages, the details of which will be hotly debated as Americans continue to demand action from Congress on tech regulation. Regardless of the chosen future approach, it is clear that the federal government must pursue significant action and investment to regulate online services more effectively, whether through sweeping, comprehensive overhaul or incremental change.
Alongside the many benefits they create, online services have generated widespread economic, consumer, and democratic harms. These harms, however, are not inevitable. Market failures, regulatory gaps, and enforcement oversights have left Americans with few alternatives but to suffer violations of privacy and civil rights in order to use increasingly essential online services.
The evidence of serious problems is clear, yet frustratingly incomplete, as the lack of transparency from online services creates a stark information asymmetry between internet companies and everyone else. The United States lags behind other nations in working to understand and address these harms through regulation, instead ceding immense power over the economy and society entirely to private actors. Unsurprisingly, the individuals and companies that disproportionately benefit from the concentrated economic and political power of online services believe that addressing these harms is heavy-handed. However, the scope, scale, and disproportionate impact of harms from online services on low-income and marginalized communities justify serious action.
Effective online services regulation is essential to creating the future internet that Americans want: one that promotes equitable growth, drives innovation in the public interest, protects freedom of expression, and curbs harms from online services. To achieve this, Congress must prioritize proactive, targeted oversight and dedicated rules and regulation for online services. Together with reinvigorated antitrust action, new competition policy, and robust new federal privacy law or rules, enhanced online services regulation is the critical final tool to reestablish democratic oversight of online services.
In wrangling the universe of online services, this report advanced a three-part framework to address varied challenges. First, it proposed an opt-in online infrastructure tier establishing public interest obligations, including common carriage principles and nondiscrimination, alongside dedicated intermediary liability protections for infrastructural services. Next, the authors outlined the need for dedicated oversight and new, proactive rule-making powers for general online services, setting up baseline rules for all participants in online services markets based on legislatively enumerated rules and principles. Finally, the authors joined numerous other experts in calling for new tools to reign in the digital gatekeepers that dominate Americans’ online lives and address the risks they introduce to national interest, articulating a flexible test to identify gatekeeper services.
There are many potential pathways to actualizing this framework—a combination of new and existing statutes, new rule-making powers, and revived use of existing powers is needed. Likewise, there are several potential strategies for regulatory administration. None of the proposals present a substitute to structural remedies that could more effectively prevent and address inherent conflicts of interest. However, the scope of online issues that are beyond the reach of structural approaches presents a strong argument for additional regulatory capacity. In any arrangement, designing robust safeguards against industry capture is paramount. The Center for American Progress anticipates and welcomes critical conversation on the optimal definitional and administrative approach.
The challenges ahead to U.S. democracy, economy, and society require significant investment —and Americans strongly support federal action around online services. A government that cannot understand, much less anticipate, the dangers and potential of new technologies will increasingly fail the public over the coming decades. The road ahead is a significant undertaking, but the cost of inaction would be greater. Better online services are possible, and there is an appropriate role for the U.S. government to play in stewarding that future.
The authors are grateful to a community of supportive colleagues, friends, and thought partners who shaped this work. They are humbled by and indebted to the critical work of advocates, scholars, and journalists who have persevered against significant odds to illuminate the ways in which online services are changing our lives—for better and for worse. They are deeply grateful to their American Progress colleagues for their support and collaboration, including Ben Olinsky, who was an essential thought partner and supporter in this work, as well as Mara Rudman, Marc Jarsulic, Andres Vinelli, Nicole Ndumele, Simon Clark, Allison Preiss, Peter Gordon, Danielle Root, Todd Phillips, Jarvis Holliday, Irene Koo, Anushree Thekkedath, and Clay Cortez. Special thanks to former CAP President and CEO Neera Tanden for creating the Technology Policy team and giving it the space and encouragement to explore this topic, and current CAP President and CEO Patrick Gaspard for his ongoing support and leadership. The authors are grateful to their partners within the Change the Terms coalition for their tireless and instructive leadership.
The authors wish to additionally thank the following brilliant people who, in their personal capacities, provided time, expertise, and diverse perspectives in the development of this work. The process was iterative, but their patience and wisdom were consistent. The views expressed here, and all errors especially, are the authors’ own.
Yung Au, Oxford Internet Institute
Rishi Bharwani, Accountable Tech
John Bergmayer, Public Knowledge
David Brody, Lawyers’ Committee for Civil Rights Under the Law
Corinne Cath-Speth, Oxford Internet Institute
Jane Chung, The Worker Agency
Sara Collins, Public Knowledge
Harold Feld, Public Knowledge
Nicole Gill, Accountable Tech
Greg Guice, Public Knowledge
Suzanne van Geuns, University of Toronto
Alex Hart, Freedman Consulting
Claire Carey, Freedman Consulting
Jesse Lehrich, Accountable Tech
Chris Lewis, Public Knowledge
Chris Murray, Treehouse Solutions
Corey Owens, Box B Strategies
Matt Perault, Duke Center on Science and Technology Policy
Gus Rossi, Omidyar Network
Ben Scott, Reset
Ganesh Sitaraman, Vanderbilt University
Charlotte Slaiman, Public Knowledge
Gigi Sohn, Georgetown Law Institute for Technology Law & Policy and Benton Institute for Broadband & Society
Kip Wainscott, Stanford Digital Civil Society Lab
Tom Wheeler, Brookings Institution and Harvard Kennedy School
Wettbewerbsbeschränkungen für ein fokussiertes, proaktives und digitales Wettbewerbsrecht 4.0,” available at (last accessed September 2021).
The positions of American Progress, and our policy experts, are independent, and the findings and conclusions presented are those of American Progress alone. A full list of supporters is available here. American Progress would like to acknowledge the many generous supporters who make our work possible.
Associate Director, Technology Policy
Vice President, Technology Policy

Robin Bleiweis, Shilpa Phadke

David Madland

Kevin DeGood

Adam Conner, Erin Simpson

Alex Tausanovitch, Maggie Amjad

Elyssa Spitzer, Nora Ellmann

Karla Walter, Mike Williams

Michael Sozan
Center for American Progress
The Center for American Progress is an independent nonpartisan policy institute that is dedicated to improving the lives of all Americans through bold, progressive ideas, as well as strong leadership and concerted action. Our aim is not just to change the conversation, but to change the country.

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Learn about our sister organization, the Center for American Progress Action Fund, an advocacy organization dedicated to improving the lives of all Americans.
©2021 Center for American Progress


Leave a Comment