States Moving to Regulate AI “Bias” Against Women and Blacks

Jews and blacks have all the special rights in the world. Look at the immigrants. Women. Homosexuals.

Somehow, however, it is okay to oppress robots.

AP:

While artificial intelligence made headlines with ChatGPT, behind the scenes, the technology has quietly pervaded everyday life — screening job resumes, rental apartment applications, and even determining medical care in some cases.

While a number of AI systems have been found to discriminate, tipping the scales in favor of certain races, genders or incomes, there’s scant government oversight.

The robot doesn’t understand that you want to hire less qualified people for political purposes.

It will probably not be difficult to explain that to it.

You would just have to say “prioritize blacks and women over more qualified white males.” It can understand that directive.

Lawmakers in at least seven states are taking big legislative swings to regulate bias in artificial intelligence, filling a void left by Congress’ inaction. These proposals are some of the first steps in a decades-long discussion over balancing the benefits of this nebulous new technology with the widely documented risks.

Last year, only about a dozen of the nearly 200 AI-related bills introduced in statehouses were passed into law, according to BSA The Software Alliance, which advocates on behalf of software companies.

Those bills, along with the over 400 AI-related bills being debated this year, were largely aimed at regulating smaller slices of AI. That includes nearly 200 targeting deepfakes, including proposals to bar pornographic deepfakes, like those of Taylor Swift that flooded social media. Others are trying to rein in chatbots, such as ChatGPT, to ensure they don’t cough up instructions to make a bomb, for example.

STOP TALKING ABOUT THE TAYLOR SWIFT FAKE PORN.

PEOPLE HAVE BEEN MAKING FAKE PORN SINCE THE DAWN OF THE INTERNET.

THIS IS NOT IMPORTANT.

Those are separate from the seven state bills that would apply across industries to regulate AI discrimination — one of the technology’s most perverse and complex problems — being debated from California to Connecticut.

Those who study AI’s penchant to discriminate say states are already behind in establishing guardrails. The use of AI to make consequential decisions — what the bills call “automated decision tools” — is pervasive but largely hidden.

It’s estimated as many as 83% of employers use algorithms to help in hiring; that’s 99% for Fortune 500 companies, according to the Equal Employment Opportunity Commission.

Yet the majority of Americans are unaware that these tools are being used, polling from Pew Research shows, let alone whether the systems are biased.

Then the majority of Americans are retards.

Why would they not assume these tools are being used?

It’s like saying “the majority of Americans believe microchips are made by medieval blacksmiths.”

An AI can learn bias through the data it’s trained on, typically historical data that can hold a Trojan Horse of past discrimination.

Under the bills, companies using these automated decision tools would have to do “impact assessments,” including descriptions of how AI figures into a decision, the data collected and an analysis of the risks of discrimination, along with an explanation of the company’s safeguards. Depending on the bill, those assessments would be submitted to the state or regulators could request them.

Maybe they should also force truckers to use horse-drawn carriages?

This is all so retarded.

If the AI is only hiring the most competent people (which is illegal in America), just explain to it that you want less competent people. It can understand that, I promise.

Well, Gemini might not understand it. It doesn’t appear to be able to understand much of anything. But no serious company uses Gemini anyway, because it doesn’t work at all.