Citizen Portal

Lawmakers press for audits, transparency and civil‑rights protections as AI spreads in schools and hiring

House Committee on Education and Workforce · January 15, 2026

Get AI-powered insights, summaries, and transcripts

Subscribe
AI-Generated Content: All content on this page was generated by AI to highlight key points from the meeting. For complete details and context, we recommend watching the full video. so we can fix them.

Summary

Committee members and witnesses urged federal leadership and stronger auditing, transparency and enforcement to prevent discrimination by hiring algorithms and protect students from privacy and safety harms; some guidance from agencies was described as removed from public websites.

Lawmakers used the House Committee on Education and Workforce hearing to highlight legal and civil‑rights risks posed by rapid AI adoption in schools and hiring tools, pressing witnesses and each other for policy remedies.

Alexandra Reeve Givens of the Center for Democracy & Technology warned that algorithmic tools in hiring and student monitoring can reproduce historical biases and reduce visibility into discriminatory outcomes. She and other witnesses said guidance from agencies such as the Equal Employment Opportunity Commission and Department of Education that once helped employers and schools comply with laws has been removed from agency websites, and they urged Congress to restore or replace that guidance and to strengthen auditing and disclosure rules for vendors.

Members repeatedly asked whether existing laws require ongoing audits; Givens said current enforcement depends on employers' liability but that discovery is difficult without transparency. Representative Summer Lee and others cited a Common Sense Media study alleging bias in a behavior‑intervention product; the vendor disputed the study and said it could not replicate findings but sought the underlying data. Lawmakers referenced pending bills such as the No Robot Bosses Act (HR6371) and said they would reintroduce or support legislation to require disclosures and testing of automated decision systems.

The hearing also focused on student safety: witnesses described the rise of AI‑generated nonconsensual imagery and deepfakes, noted the increased volume of tips to the National Center for Missing and Exploited Children, and urged companies to improve detection, takedown, and reporting practices. Members asked for clearer pathways for victims to remove content, and cited recent legislative steps described by witnesses as the 'take it down' act and related measures.

Lawmakers asked the panel to produce additional technical details and welcomed further oversight; no regulatory action was taken at the hearing, but members signaled intent to pursue legislative and oversight options.