Read the full report
This fact sheet collects the recommendations from Chapter 2: “The Department of Labor” of the joint report from Governing for Impact (GFI) and the Center for American Progress, “Taking Further Agency Action on AI: How Agencies Can Deploy Existing Statutory Authorities To Regulate Artificial Intelligence.” The chapter notes how the U.S. Department of Labor (DOL) oversees numerous statutes, from the Fair Labor Standards Act (FLSA) to the Family and Medical Leave Act (FMLA), that can potentially help address the challenges and opportunities of artificial intelligence (AI) as it affects workers. These recommendations stem from DOL-enforced statutes identified in the chapter that could be used to address AI through regulations, subregulatory guidance, and enforcement practices. Among other authorities, the DOL could use these statutes to ameliorate known harms by updating wage and hour regulations, guarding workers’ safety and health against the negative impacts of automated management, and ensuring that automated benefits administration is transparent and fair. The goal of these recommendations is to provoke a generative discussion about the following proposals, rather than outline a definitive executive action agenda. This menu of potential recommendations demonstrates that there are more options for agencies to explore beyond their current work, and that agencies should immediately utilize existing authorities to address AI.
Read the full chapter
Fair Labor Standards Act: Recordkeeping and reporting
Based on this authority, the DOL could consider the following actions:
- Issue new recordkeeping and reporting rules, pursuant to 29 U.S.C. § 211(c), to require employer records to ensure legibility and transparency of wage determinations made by automated systems and to require periodic reports to the Wage and Hour Division (WHD) of those records from employers using AI-driven wage and scheduling technology. Such regulations would help combat black-box wage determination and discrimination1 that can make workers’ wages unpredictable and irregular,2 as well as ensure that such wage determinations satisfy the minimum wage and overtime requirements of the FLSA. As documented by Veena Dubal, professor of law at the University of California, Irvine, many workers are subject to algorithmic management and wage setting that withholds or reduces compensation for work when doing so benefits the company.3 This can make it difficult for workers to appreciate the connection between time spent working and amount of income generated, or to understand and correct errors in their compensation, and can also result in opaque wage setting that violates minimum wage or overtime laws.4 The DOL contemplated a similar rulemaking in the early 2010s that would have required recordkeeping and disclosure to workers about their status as employees or independent contractors and detailed information about how their pay is computed, but a regulation was never proposed.5
- Launch investigations, pursuant to its administrative subpoena power in 29 U.S.C. § 211(a),6 of employers to ensure compliance with minimum wage and overtime provisions. The WHD could prioritize investigation of employers that are noncompliant with the reporting rules mentioned, are in industries with large numbers of employee complaints, or are in industries with high penetration of automated wage and scheduling technologies. These investigations could produce valuable information about the characteristics of automated systems that make minimum wage and overtime violations more likely to occur and encourage employers’ compliance with their legal obligations under the FLSA.
Fair Labor Standards Act: Minimum wage and overtime
Based on the above-cited authority, the DOL could consider the following action:
- Issue updated interpretive regulations at 29 C.F.R. Part 785, pursuant to 29 U.S.C. § 211(c), that allow only employers who track time manually through analog methods to engage in timesheet rounding7 and establish a presumption against application of the de minimis rule in cases where employers use highly precise timekeeping technology.8 These changes would eliminate an outdated regulatory regime that allows companies to use sophisticated timekeeping technology to facilitate wage theft by exploiting rules meant to minimize the burden of pen-and-paper wage and hour calculations. Given the ubiquity and ease of digital timekeeping, there is no longer a compelling justification for allowing practices such as rounding employees’ hours to the nearest quarter-hour or failing to treat short periods of working time as compensable for minimum wage and overtime compliance.9
Unemployment compensation
Based on the above-cited authority, the DOL could consider the following actions:
- Update quality control program regulations at 20 CFR § 602.21, pursuant to 42 U.S.C. §§ 503(a)(1) and 1302, to require states to undertake audits and submit their results to the DOL for any automated or AI-driven benefits determination system. This could help ensure that states provide unemployment compensation to individuals consistent with federal law, provide for human in-the-loop review of any algorithmic denial of benefits, and ensure fair human adjudication for appeals of those denials. The current quality control program regulations were promulgated based on this same statutory authority.10 These regulations would guard against states’ use of automated systems to deny coverage to eligible individuals (or worse, wrongfully accuse them of fraud),11 a use case cited by the Office of Management and Budget (OMB) as presumptively rights-impacting, and therefore it should be subject to heightened scrutiny.12 This proposal is closely related to the actions directed in Section 7.2(b) of the president’s 2023 executive order on AI, which aims to ensure the equitable distribution of public benefits. For example, the executive order directs the U.S. Department of Agriculture to issue guidance to state, local, and Tribal governments that address the use of AI systems in benefits distribution. It requires such guidance to ensure that such systems, among other things, maximize program access; require governments to notify the Department of Agriculture of AI use; create opt-out opportunities for benefit denial appeal; and enable auditing to ensure equitable outcomes.13
- Issue a new unemployment insurance program letter (UIPL) to guide states specifically on where and how AI can and should be implemented for unemployment insurance administration. This new UIPL should incorporate the minimum risk management practices for the presumed rights-impacting use of AI from theOMB M-24-10 AI memo14 and any subsequent guidance. For example, utilizing AI to flag potential fraud must be accompanied by the minimum risk practices from the OMB M-24-10 AI memo, such as carrying out AI impact assessments, testing the systems in the real world before widespread deployment, and ongoing monitoring to ensure equity.15 The DOL should clarify that these requirements extend to any vendor a state unemployment insurance system contracts with to provide services.
Occupational Safety and Health Act
Based on the above-cited authority, the DOL could consider the following actions:
- Begin the standard-setting process, pursuant to 29 U.S.C. § 655(b), to regulate the use of electronic surveillance and automated management (ESAM) in the workplace to the extent that it creates hazards to workers’ physical and mental safety and health. Such regulation could mitigate the increasingly unsustainable pace of work enforced by these systems, which leads to ergonomic injury and increased risk of accidents. For example, the Washington State Department of Labor and Industries has fined Amazon repeatedly for forcing its warehouse workers to work at punishing speeds that exacerbate the risk of injury.16 The state’s citations specifically reference the “direct connection” between Amazon’s ESAM and workplace musculoskeletal disorders.17 A standard on ESAM would also reduce the harmful effects that these systems can have on workers’ mental health. As early as 1987, the now-defunct U.S. Office of Technology Assessment recognized that ESAM increases employee stress, heightening job strain risk.18
Of course, the Occupational Safety and Health Administration’s (OSHA) standard-setting process is uniquely slow and resource intensive for the agency,19 and the process would need to be informed by additional research to design an effective policy. So, in the meantime, the following recommendations should be considered:
- Issue new subregulatory guidance and bring general duty clause enforcement actions related to companies’ use of ESAM in ways that harm worker safety and health. As GFI has urged in past advocacy efforts, OSHA should follow the lead of Washington state by more directly tying ESAM use to physical and mental health hazards.20 Enforcement actions based on unsafe ESAM use could be taken because of the already ongoing DOL investigation into high injury rates at Amazon warehouses.21
- Update existing subregulatory guidance about sector-specific ergonomic risks to include a discussion of how ESAM can increase musculoskeletal injury risk. As described in a GFI report in 2023, OSHA could update the ergonomics guidance documents for poultry processing and grocery warehousing and create a new ESAM-conscious ergonomic risks guidance document for the warehousing industry.22 The guidance could describe best practices to prevent ergonomic injuries—such as quota transparency, worker involvement in quota setting, and rest breaks—and how ESAM systems should be adjusted to accommodate those best practices.
- Update injury reporting regulations at 29 C.F.R. Part 1904, pursuant to 29 U.S.C. § 657, revising OSHA’s log of work-related injuries and illnesses (Form 300) to collect information about automated systems used in the tasks, job roles, or workplaces in which the worker was working at the time of injury or illness. Additionally, OSHA could update Form 300 to include a column identifying when injuries are musculoskeletal.23 This would allow OSHA to develop a better understanding of the precise causal mechanisms between ESAM and these injuries and inform the substantive policymaking described above.
- Request research from the National Institute for Occupational Safety and Health, pursuant to 29 U.S.C. § 671(d), to fund and conduct further research to study ESAM’s effect on job strain and physical injury.24
Employee Retirement Income Security Act: Adverse benefits determination and disclosure
Based on the above-cited authority, the DOL could consider the following actions:
- Update regulations at 29 C.F.R. § 2560.503-1, which implement the denial-of-claims disclosure and appeal requirements at 29 U.S.C. § 1133. The current regulations state, for example, that in the case of an adverse benefit determination by a group health plan, a participant is entitled to request a copy of any “internal rule, guideline, protocol, or other similar criterion” that was relied on in making the adverse determination.25 An updated regulation could require affirmative disclosure of a plain-language description of any algorithmic determination involved in a benefits determination, as well as the results of an equity audit conducted in a manner similar to that recommended in the OMB M-24-10 AI memo.26 Additionally, the updated regulations could clarify that the appeal process authorized by 29 U.S.C. § 1133(2) and outlined at 29 C.F.R. § 2560.503-1(h) requires that appeals of benefits denials be heard by a human. This update could come as part of the DOL’s announced review of the Employee Retirement Income Security Act (ERISA) disclosures pursuant to the Setting Every Community Up for Retirement Enhancement (SECURE) Act 2.0.27
- Update regulations at 29 C.F.R. § 2520.102-3(l) to amend the summary of plan description to include a plain language description of any automated and algorithmic systems that the plan uses to make determinations that could “result in disqualification, ineligibility, or denial or loss of benefits,”28 as well as whether the system has been externally audited or the administrator has instituted safeguards such as opt-out mechanisms for participants who would prefer human-made determinations. This would provide some transparency to workers and advocates about the decisions that plan administrators make with the help of AI-driven systems. This update could also come as part of the DOL’s announced review of ERISA disclosures pursuant to the SECURE Act 2.0.29
Employee Retirement Income Security Act: Investment advice
Based on the above-cited authority, the DOL could consider the following actions:
- Update regulations at 29 C.F.R. § 2550.404a-1(c), pursuant to 29 U.S.C. § 1104, to revise the investment duty of loyalty in light of the risks that AI-driven investment allocation technologies can create and potential conflicts of interest. The updated regulation could be similar to the U.S. Securities and Exchange Commission’s rulemaking proceedings that seek to prevent investment advisers from using algorithms that create conflicts of interest between the adviser and the investor’s retirement goals.30 Importantly, plan fiduciaries should be required to ensure that AI-driven investment advice or allocations are not improperly weighted toward decisions that maximize fees and commissions at the expense of retirement savers. Such regulations could also require an audit of any AI-driven or otherwise automated investment allocation technologies for the potential for conflicts of interest.
- Issue new regulations, pursuant to 29 U.S.C. § 1104(c)(5), requiring algorithmic transparency and legibility to plan participants and beneficiaries for default asset allocations.31
- Update the statutory transactions exemption at 29 C.F.R. § 2550.408g-1(b)(4), “Arrangements that use computer models,” to strengthen the existing auditing requirements and institute other AI-specific requirements, taking into account the DOL’s approach in the proposed revisions to the Prohibited Transaction Exemption 2020-02.32 Alternatively, or in addition to updating the exemption, the DOL could issue guidance that more fully describes the term “computer model” and identifies AI applications to which this exemption may apply.
Labor Management Reporting and Disclosure Act
Based on the above-cited authority, the DOL could consider the following action:
- Issue a regulation or subregulatory guidance, in the form of independent guidance documents or in the LM-10 form instructions, that explains how forms of ESAM can chill workers’ exercise of their Section 7 rights under the National Labor Relations Act and when they must be reported in employers’ LM-10 forms. The use of worker surveillance to thwart organizing activities is well-documented.33 The regulation or guidance could explain how that might require employers to report their expenditures on such technologies. They could reference the memo issued by the National Labor Relations Board’s general counsel on the subject,34 as well as prior guidance from the DOL on surveillance reporting.35 Additional guidance may empower workers, unions, and labor watchdogs to report employer noncompliance to the DOL.
Worker Adjustment and Retraining Notification Act
Based on the above-cited authority, the DOL could consider the following action:
- Update regulations at 20 C.F.R. § 639.3(i), pursuant to 29 U.S.C. § 2107(a), to explain that, in the case of a completely or primarily remote workforce, the term “single site of employment” applies to the employer’s entire workforce. In the case of algorithmic management, the DOL should clarify that all workers subject to the same or similar algorithm are considered one single site of employment. Updated regulations could also ensure that workers subject to intermittent deplatforming caused by algorithmic optimization have maximal protections possible under the Worker Adjustment and Retraining Notification (WARN) Act.
Family and Medical Leave Act
Based on the above-cited authority, the DOL could consider the following actions:
- Update regulations at 29 C.F.R. Part 825, pursuant to 29 U.S.C. §§ 2615(a)(1) and 2654, to require legibility and transparency of automated systems36 that make any determinations bearing on the allocation or approval of FMLA leave, along with any other applicable minimum practices for rights-impacting AI from the OMB M-24-10 AI memo.37 This would implement the transparency protections recommended by the White House’s AI Bill of Rights and ensure that employers’ use of automated systems does not unlawfully restrain workers’ exercise of their rights under the FMLA. Because FMLA determination algorithms are likely bound up in other human resource management systems, this proposal could also provide transparency of those benefits processes as well. Specifically, these updated regulations should require:
- At 29 C.F.R. § 825.301, legibility and transparency around use of automated systems to make FMLA designations
- Legibility and transparency around use of automated systems to review, request, or otherwise process certifications under 29 U.S.C. § 2613
- Legibility and transparency around use of automated systems to provide eligibility notices, at 29 C.F.R § 825.300(b); rights and responsibilities notices, at 29 C.F.R. § 825.300(c); and designation notices, at 29 C.F.R. § 825.300(d)
- At 29 C.F.R. § 825.302, legibility and transparency around use of automated systems for employees to provide notice of the use of leave or to transmit information around scheduling of intermittent leave under 9 U.S.C. § 2612(b) and (e)
- Update regulations by modifying 29 C.F.R. § 825.220, pursuant to 29 U.S.C. §§ 2615(a)(1) and 2654, to prohibit employers from using FMLA data as inputs to any automated management system that may make an employment decision based, in part, on an employee’s use or nonuse of FMLA leave. This would reduce employers’ ability to weaponize employees’ data against them to retaliate for using FMLA leave. Under these recommended updated regulations, the automated management system must strictly segregate and keep confidential any information provided for FMLA certification pursuant to 29 C.F.R. § 825.500(g).
- Update subregulatory guidance under 29 C.F.R. § 825.301(a) prohibiting automated systems from using information other than that received from the employee or the employee’s authorized spokesperson in designating FMLA leave pursuant to 29 C.F.R. § 825.301(a). Existing regulation already prohibits the conduct for employers and would also apply to automated systems used by employers, but additional clarification is essential to restrict automated systems that would improperly combine data sources.