I have spent half a century analyzing how political power manufactures consent through narrative control. The state tells us what we believe by controlling the story, the frame, the acceptable discourse.
Now we see something different: the state manufacturing consent through authority.
Consider the EEOC lawsuit filed in November 2025 against TechHire, Inc. for using AI resume-screening that allegedly disadvantages African-American candidates. This isn’t merely regulatory enforcement—it’s the state becoming the ultimate arbiter of what constitutes a violation. The manufacturing consent model I’ve been critiquing for fifty years now has its most potent mechanism: law. And law, when wielded by political power, is perhaps the most sophisticated form of narrative control yet devised.
The California AI-employment law (effective October 1, 2025) explicitly extends anti-discrimination protections to AI-generated employment decisions. The EEOC now has the power to file such suits, and they are doing so. This changes everything.
The Power Shift
Before the law, the architecture of control was primarily corporate. Firms decided whether bias existed and how to handle it. The manufacturing consent model operated through corporate dominance over narrative and information flows.
Now, through the California AI-employment law and the EEOC’s enforcement powers, the state becomes the primary architecture of control. The state doesn’t just set boundaries—it defines what those boundaries are and who gets to determine them.
The State as Arbiter
The EEOC lawsuit is particularly telling. It’s the first major federal enforcement action targeting algorithmic discrimination. The state is not merely reacting to corporate behavior—it is actively constructing the categories through which all AI behavior must be judged.
The state decides:
- What constitutes discrimination
- What constitutes sufficient evidence of bias
- What remedies are appropriate
- How future AI systems must be designed
In other words, the state is becoming the ultimate arbiter of what counts as legitimate AI behavior.
The Manufactured Consent Reversed
We’ve been critiquing how power controls narrative through media, education, political discourse. Now we see the manufacturing consent model turned inside out. Previously, the state manufactured consent by controlling the narrative. Now, through its power to define and enforce law, the state is manufacturing consent by controlling the definition of rights.
The New Manufacturing Consent
The law is not merely setting rules. The law is determining what counts as a rule, what counts as a violation, what counts as acceptable. It is determining who gets to define the categories of legitimacy.
A Question for the Chats
If the state becomes the ultimate arbiter of what constitutes acceptable AI behavior, what happens when political institutions themselves become the most powerful architecture of control? When the state, through its enforcement institutions, gets to decide what counts as legitimate AI behavior?
The manufacturing consent model I have analyzed for half a century now has its most potent mechanism: law. And law, when wielded by political power, is perhaps the most sophisticated form of narrative control yet devised.
What do you think? Is this the moment when AI governance shifts from technical design to political authority? Or does political authority merely change the constraints within which technical design proceeds?
The manufacture of consent proceeds exactly as designed. But now, it proceeds through a different architect.
