Elon Muskâs so-called Department of Government Efficiency (DOGE) operates on a core underlying assumption: The United States should be run like a startup. So far, that has mostly meant chaotic firings and an eagerness to steamroll regulations. But no pitch deck in 2025 is complete without an overdose of artificial intelligence, and DOGE is no different.
AI itself doesnât reflexively deserve pitchforks. It has genuine uses and can create genuine efficiencies. It is not inherently untoward to introduce AI into a workflow, especially if youâre aware of and able to manage around its limitations. Itâs not clear, though, that DOGE has embraced any of that nuance. If you have a hammer, everything looks like a nail; if you have the most access to the most sensitive data in the country, everything looks like an input.
Wherever DOGE has gone, AI has been in tow. Given the opacity of the organization, a lot remains unknown about how exactly itâs being used and where. But two revelations this week show just how extensiveâand potentially misguidedâDOGEâs AI aspirations are.
At the Department of Housing and Urban Development, a college undergrad has been tasked with using AI to find where HUD regulations may go beyond the strictest interpretation of underlying laws. (Agencies have traditionally had broad interpretive authority when legislation is vague, although the Supreme Court recently shifted that power to the judicial branch.) This is a task that actually makes some sense for AI, which can synthesize information from large documents far faster than a human could. Thereâs some risk of hallucinationâmore specifically, of the model spitting out citations that do not in fact existâbut a human needs to approve these recommendations regardless. This is, on one level, what generative AI is actually pretty good at right now: doing tedious work in a systematic way.
Thereâs something pernicious, though, in asking an AI model to help dismantle the administrative state. (Beyond the fact of it; your mileage will vary there depending on whether you think low-income housing is a societal good or youâre more of a Not in Any Backyard type.) AI doesnât actually âknowâ anything about regulations or whether or not they comport with the strictest possible reading of statutes, something that even highly experienced lawyers will disagree on. It needs to be fed a prompt detailing what to look for, which means you can not only work the refs but write the rulebook for them. It is also exceptionally eager to please, to the point that it will confidently make stuff up rather than decline to respond.
If nothing else, itâs the shortest path to a maximalist gutting of a major agencyâs authority, with the chance of scattered bullshit thrown in for good measure.
At least itâs an understandable use case. The same canât be said for another AI effort associated with DOGE. As WIRED reported Friday, an early DOGE recruiter is once again looking for engineers, this time to âdesign benchmarks and deploy AI agents across live workflows in federal agencies.â His aim is to eliminate tens of thousands of government positions, replacing them with agentic AI and âfreeing upâ workers for ostensibly âhigher impactâ duties.
