Introduction

If the US federal government is unquestionably good at one thing, it is pushing out paper. In theory, governmental records relevant to the creation of policy via agency rulemaking have long been available to researchers. Given its volume and quasi-legislative nature, rulemaking is generally required to be more transparent than more individualized decision-making such as adjudication, enforcement, awarding grants, or making contracts. In practice, however, obtaining information on what is happening during rulemaking has been costly and challenging. In the 1990s and early 2000s, leading research was limited to data on just a few rules (e.g., Golden 1998) or surveys (e.g., Furlong 2004). Since 1994, when the government first released the Federal Register online, data on notice-and-comment rulemaking have become increasingly detailed and now include data on draft policies, the activities of policymakers, and interest group advocacy (see Yackee 2019 for a recent review).

In this article, we describe data sources that have proven useful to scholars studying interest group lobbying of federal agencies, how to obtain them, and also challenges in working with these data. Some of these data sources we discuss are machine-readable records of all agency rules published in the Federal Register, comments posted on regulations.gov, metadata about rules contained in the Unified Agenda, and Office of Management and Budget (OMB) reports. We also describe sources that have become more available in recent years, such as ex-parte meeting logs and individually identified personnel records of nearly all federal employees since 1973. For background theory on lobbying and interest group influence over agency policymaking, see Carpenter and Moss (2013).

These data sources do not exhaust the kinds of records relevant to researchers. They reflect what is available at present, but new data sources emerge constantly. Thus, we also highlight datasets that may soon become more accessible or available, perhaps after enterprising researchers submit the necessary Freedom Of Information Act (FOIA) requests. Examples include agency press releases (see, e.g., Libgober 2019; Libgober and Carpenter 2018) and the Foreign Agents Registration Act (FARA) reports (Shepherd and You 2019).Footnote 1 To orient potential students of agency policymaking to available sources of data, we identify four units of analysis: (1) participants, (2) policymakers, (3) policy texts, and (4) metadata (e.g., policy timing). We describe where to find data on each in roughly the order that they appear in the process of developing a rule.

Background: rulemaking and the Administrative Procedures Act

For decades, the volume of legal requirements emerging from executive agencies has dwarfed the lawmaking activity of Congress. Each year, agencies publish three thousand or more regulations. In doing so, agencies must generally follow a process prescribed by the Administrative Procedures Act (APA), 5 U.S.C. §553(c). For each rule, agencies create a collection of documents at each stage of their policymaking process. Section 553 requires agencies to issue a Notice of Proposed Rulemaking (NPRM) in the Federal Register to notify potentially affected parties that the regulatory environment or program administration may change. Following the NPRM, agencies post these proposed rules and comments from interested stakeholders and individuals on their websites or regulations.gov. Agencies are required to consider public comments but are not required to alter rules based on them. These documents are organized in a ‘docket’ folder with a unique name and ID number. Once kept in literal folders in each agency’s ‘docket room,’ where visitors could make photocopies, these documents are now online. For example, the Centers for Medicare & Medicaid Services rulemaking docket on health plans disclosing costs can be accessed at https://www.regulations.gov/docket?D=CMS-2019-0163. The fact that the APA requires agencies to assemble a comprehensive record of who sought to influence the policymaking process—and that such records are relatively complete—is unusual for federal policymaking. This makes agency policymaking particularly exciting for interest group scholars.

Who participates in the rulemaking process?

Public attention in rulemaking is remarkably skewed. A few rules, such as the Federal Communication Commission’s rules on net neutrality, receive millions of comments. In contrast, half of the proposed rules open for comment on https://www.regulations.gov/ received no comments at all (Libgober 2020). Such dramatic variation in public participation challenges the idea of a “typical” rule or rulemaking participant. Participants include businesses, public interest groups, trade associations, unions, law firms, and academics (Cuéllar 2005; Yackee and Yackee 2006). While businesses are the most consistent and influential participants in most rulemakings (Yackee and Yackee 2006; Libgober 2020), most comments come from public pressure campaigns targeting a relatively small number of rules; at least 39 million of the 48 million comments on proposed rules on regulations.gov were mobilized by just 100 advocacy organizations such as the Sierra Club (Judge-Lord 2019).

Because interest groups seek to influence rules, scholars are interested in patterns of participation in rulemaking. For example, You (2017) finds that half of all spending on lobbying legislation occurs after a bill becomes law. Participation in this process can begin even before a proposed rule is issued. de Figureido and Kim (2004) show that meetings between agency officials and firms spike before an agency issues a policy. Libgober (2019) finds that firms that meet with federal regulators before a rule is issued may receive abnormally high stock market returns upon its release. Leveraging high-frequency trading data, Libgober shows that in the minutes and hours following the publication of proposed rules at the Federal Reserve, the firms that met with the Board during rule-development significantly outperformed matched market competitors that did not obtain such early access. These findings are consistent with the analysis of qualitative researchers that commenters who participate early in the rulemaking process can shape the content of the rule (Naughton et al. 2009).

Scholars are also interested in how participation in rulemaking may affect a rule’s ultimate fate when it is sent to the White House Office of Management and Budget or in judicial review. Interest groups can secure policy concessions by lobbying during OMB review (Haeder and Yackee 2015). They also use the rulemaking record to build a case for litigation. Yet, there are few large-N studies linking lobbying in rulemaking to litigation or court decisions. Libgober and Rashin (2018) analyzed comments submitted to financial regulators and found that threats of litigation were rare, and even comments that seemed to “threaten” litigation rarely culminated in a judicial decision. Yet, most court decisions involving challenges to rules did follow comments that threatened litigation. Judge-Lord (2016) found no relationship between the number of comments and the likelihood that the Supreme Court upheld or struck down an agency rule.

To analyze patterns of participation, scholars use data from a variety of sources. Sources for data on participation are agency rulemaking dockets (Golden 1998; Yackee 2006; Young et al. 2017; Ban and You 2019), the Federal Register (Balla 1998; West 2004), and regulations.gov (Balla et al. 2019; Gordon and Rashin nd). Though commenters are not generally required to disclose their names and affiliations, many do. The best current data sources for obtaining the names of the organizations that submit comments on federal regulations are regulations.gov and the websites of the independent agencies themselves.

Obtaining and working with data on comment participants

Scholars wishing to obtain data on comments and commenters from executive agencies generally use the website regulations.gov. Obtaining these data requires overcoming several technical and bureaucratic hurdles. First, retrieving bulk data from regulations.gov requires the use of an Application Programming Interface (API).Footnote 2 Second, getting an accurate count of comments is not straightforward as agencies have different policies regarding duplicated commentsFootnote 3 and confidential business information (Lubbers 2012).

Third, downloading these data can be time-consuming. As of February 29, 2020, Regulations.gov has 12,227,522 public submission documents representing over 70 million public comments, over 80,000 rules, and nearly 1.5 million other documents. Regulations.gov is subject to a rate limit of 1000 queries per hour, making it difficult for scholars seeking to analyze rules with tens of thousands of comments. This limit is particularly problematic for rules where the majority of comments are attachments, as downloading an individual document requires calling the API twice (once for the docket information, which includes the attachment URL(s), and then a second time to download the linked file). Fourth, not all federal agencies post rulemaking documents to regulations.gov. For these agencies, scholars studying participation in rulemaking can often obtain data from the agency’s website.Footnote 4

After obtaining comment data, scholars must choose how they want to preprocess the names of the organizations that comment. This step is vital as organizations such as Goldman Sachs and the American Bar Association, for example, submit comments using multiple versions of their names. These decisions can have a substantial impact on the analysis as misidentifying organizations may result in them being dropped or double-counted. Fuzzy matching, using an algorithm to identify the similarity (or “distance”) between two strings of text, can help but is not a panacea. For example, the algorithm will show a small distance between Goldman Sachs and Goldman Sachs & Co but can show similarly small distances between distinct entities and it cannot separate multiple organizations with the same acronym (e.g., the American Bankers’ Association and American Bar Association both comment as the ABA).

Who writes rules?

Political scientists have long used personal identity and social networks to understand political outcomes. In the last decade, the emergence of new data sources made it easier to learn about rulewriters’ identities and networks. Some of the most important data sources in this regard are the US Office of Personnel Management’s (OPM) data on government employees (e.g., Bolton et al. 2019),Footnote 5 Open Secrets’ lobbying (e.g., Baumgartner et al. 2009) and revolving door databases (e.g., i Vidal et al. 2012; Bertrand et al. 2014), machine-readable lobbying disclosure act reports (e.g., Boehmke et al. 2013; You 2017; Dwidar 2019), meeting logs (Libgober 2019), and datasets of corporate board membership such as Boardex (e.g., Shive and Forster 2016). Carrigan and Mills (2019) illustrate the possibilities of these new data sources. They find that the number of job functions of the bureaucrats who write the rules is associated with both decreases in the time an agency takes to promulgate a rule and increases in the probability that the rule will be struck down in court. In an innovative study of unionization, Chen and Johnson (2015) link OPM data to Adam Bonica’s DIME ideology scores to classify the ideology of agency staff over time. These early efforts suggest that there are exciting opportunities to study who writes the rules and how the identities of the rulewriters affect policy. 

Obtaining and working with personnel and lobbying data

Scholars can obtain data on agency personnel from the BuzzFeed personnel data release, which contains information such as employee salaries, job titles, and demographic data from 1973 through 2016.Footnote 6 These are the most comprehensive personnel records publicly accessible, but they have limitations. First, not all agencies and occupations are a part of this release.Footnote 7 Second, some of the employees in these data do not have unique identification numbers, and common names such as ‘John Smith’ match multiple employees. For example, the Veterans Health Administration employed 24 John Smiths in 2014, five with the same middle initial.

To obtain machine-readable data on the identities of domestic lobbyists, scholars use two databases from Open Secrets—a nonprofit that tracks spending on politics—on administrative and Congressional lobbying. Downloading the lobbying data is straightforward; it only requires an account to access the ‘bulk data’ page. The lobbying data comes from the required disclosures under the Lobbying Disclosure Act of 1995 (LDA). Note that the reporting threshold varies by type of firm (in-house have a higher minimum reporting threshold than lobbying firms) and over time.Footnote 8 The lobbying data cover 1999 through 2018 and are broken up into seven machine-readable tables.Footnote 9 We prefer the Open Secrets data to the raw Senate data as the Open Secrets version contains more machine-readable information. Lobbying data are subject to several limitations; the reports do not contain exact monetary amounts and some lobbyists do not disclose required contacts.Footnote 10 Open Secrets also has a database on revolving door employees that shows the career paths of federal government workers that went to private sector work that depends on interacting with the federal government.Footnote 11

Lobbyists advocating for foreign clients are required to disclose these contacts under the Foreign Agent Registration Act. Foreign agents often lobby bureaucratic agencies; Israel, for example, retained law firm Arnold & Porter for advice on registering securities with the Securities and Exchange Commission (SEC).Footnote 12 In addition to the participants, the reports also contain the nature of the contact and the specific officials contacted. These data are astonishingly complete—the website contains all 6264 FARA registrants and their foreign contacts since 1942.Footnote 13 The FARA data are not all in machine-readable form.Footnote 14 Consequently, scholars using these data only focus on a subset such as You (2019) who focuses on lobbying activities by Colombia, Panama, and South Korea. We caution users that the same participant may appear under multiple names.

Corporate executives, lawyers, and lobbyists often have contacts with agency officials, called ex-parte communications (Lubbers 2012), to discuss rules outside of the public comment process. While record-keeping practices for these meetings vary considerably, they are often publicly available on agency websites. Because meetings data are held in different places on each agency’s website, obtaining these data requires writing a web scraper for each agency. As there are no uniform standards for reporting meeting data, these data differ substantially from agency to agency in both content and organization. The Federal Reserve, for example, groups meetings by subject, not by rule,Footnote 15 the SEC groups meetings by rule, and the Commodity Futures Trading Commission (CFTC) and Federal Communication Commission (FCC) meeting records are not grouped at all.Footnote 16 Meeting participants are typically individually identified, but some agencies only report the names of the organization represented, and others report meeting topics but not participants. We suspect that missingness varies by agency, but are not aware of any study evaluating missingness in meeting record disclosures.

What do the rules say?

The literature on rulemaking in political science and public administration has focused on whether public participation influences the content of rules (e.g., Balla 1998; Cuéllar 2005; Yackee and Yackee 2006; Naughton et al. 2009; Wagner and Peters 2011; Haeder and Yackee 2015, 2018; Gordon and Rashin nd; Rashin 2018). Knowing what the rules say and how these texts have changed from proposal to finalization is crucial studying commenter influence. Rule preambles, which describe what the agency has tried to accomplish and how it has engaged with stakeholders, are especially useful. The legally operative text is often less useful to social scientists because interpreting legal text usually requires substantial domain expertise. However, where interpretation is straightforward, such as rate-setting, scholars can capture variation in legally operative texts (Balla 1998; Gordon and Rashin nd).

Obtaining and working with data on rule text

The federal government publishes all rules in the Federal Register, accessible via FederalRegister.gov. This database is comprehensive and has machine-readable records of all regulations published after 1994 and PDF versions from 1939–1993.Footnote 17 The Federal Register API does not have any rate limits. The Federal Register office assigns every regulatory action a document number. In our experience, this ‘FR Doc Number’ is the only truly unique and consistently maintained identifier for proposed and final rules.

While the raw text of rules is relatively straightforward to obtain, there are several thorny theoretical and methodological issues that scholars must overcome.Footnote 18 First, the standard path from NPRM to public comments to final rule is not always straightforward. Some rules are withdrawn before a final rule. Other agencies issue interim final rules subject change. For studies that seek to compare the proposed and final rules, the most problematic rules are the ones where one proposed rule gets broken up into a few smaller final rules or the reverse, where short rules become bundled into one final rule. These rules pose challenges for inference as the processes that lead to amalgamation or separation are not well understood. Second, scholars must determine how to preprocess the rulemaking data before feeding these data to a text analysis algorithm.

Obtaining and working with rule metadata

Rulemaking metadata, such as the time rules are released, allow scholars to answer questions about factors that affect the rulemaking environment. Scholars use this data to study questions about regulatory delay (e.g., Acs and Cameron 2013; Thrower 2018; Potter 2017; Carrigan and Mills 2019; Carpenter et al. 2011), agenda-setting (e.g., Coglianese and Walters 2016), and the financial impact of rules (e.g., Libgober 2019). These scholars relied on data from the Office of Information and Regulatory Affairs (OIRA) (Acs and Cameron 2013; Carrigan and Mills 2019), the Unified Agenda (Coglianese and Walters 2016; Potter 2017), the Federal Register (Thrower 2018), press releases (Libgober and Carpenter 2018), and the Food and Drug Administration’s drug approval and postmarket experience database (Carpenter et al. 2011).

Obtaining data from the Unified Agenda is relatively straightforward as all of these documents since 1995 are available online in machine-readable form.Footnote 19 The Unified Agenda contains many of the proposed regulations that agencies plan to issue in the near future, making it an extremely useful data source for studying questions about agenda setting and timing (e.g., Potter 2019). There are, however, significant limitations to these data. First, agencies report early stage rulemaking to the Unified Agenda strategically (Nou and Stiglitz 2016). Second, agencies do not list all ‘failed’ rules that did not become final rules in the Unified Agenda (Yackee and Yackee 2012). Third, Coglianese and Walters (2016) note that the Unified Agenda misses much of the regulatory agency’s work, including enforcement actions, adjudicatory actions, and decisions not to act.

As addressed elsewhere in this issue (Haeder and Yackee 2020 in this issue), many rules, especially rules that are controversial or deemed economically significant, are reviewed by OIRA (a subagency of OMB). Obtaining OIRA data is relatively straightforward, as all of these documents since 1981 are available online in machine-readable form. These data include the date on which OIRA received the draft rule from the agency, whether the review was expedited, and whether the rule is determined to be ‘economically significant’ or affect ‘federalism.’

In addition to disclosures mandated by law, agencies often issue press releases for agency actions. Much like the meetings data discussed above, policies regarding the storage and dissemination of press releases differ from agency to agency. The Federal Reserve’s website, for example, lists all press releases since 1996, while the SEC only has them since 2012.Footnote 20 When working with press releases, scholars often need data on the exact time documents were made available to the general public (see, e.g., Libgober 2020). Press release metadata, such as the exact time a press release becomes public, can often be extracted from Really Simple Syndication (RSS) feeds.

Each data source has different types of missingness and different limitations regarding the information it contains about each rule. For example, data from OIRA only contains rules reviewed by OMB. The Unified Agenda covers a broader scope of policies, but due to strategic reporting and frequent reporting errors, desired cases may be missing. For published draft and final rules, these missing cases may be found in a more reliable source like the Federal Register, OIRA reports, or regulations.gov. The Federal Register is the most reliable source for rule texts but contains the least amount of rule metadata and only includes published draft and final rules. Some rulemaking projects that did not reach the published draft (NPRM) stage may be missing from all datasets. More importantly, the diversity of these data sources means that, for a given query, one source will often include cases that a second does not, while the second source includes variables that first does not.

Discussion: assembling complete databases

The US government releases troves of data on rulemaking. Yet, these data require substantial effort from scholars to be useful for research. Scholars working on bureaucratic politics face two primary data tasks: (1) assembling complete, machine-readable datasets of agency rulemaking activity and (2) linking observations across datasets. In the near term, we see four major projects to make data more accessible, prevent duplicated efforts to download and clean data, and thus increase the efficiency of research efforts. First, a complete database is needed to link commenting activity throughout the Federal government. Researchers should be able to query and download these data in bulk, including comment text and metadata. Second, a comprehensive database could link observations of the revolving door for rulewriters and participants across OPM personnel records, LDA disclosure forms, and FARA data. This database could be augmented with data from networking sites like LinkedIn. Third, a database could link all meeting activities throughout the federal government. Finally, creating unique identifiers for each commenter would allow researchers to link commenting behavior across datasets and to other information about these individuals and organizations. These projects would allow scholars to pursue novel research on political participation, influence, and public management.