Is an Executive Order on AI Enough?

(Source)

 

President Biden’s Executive Order 14110

President Biden signed an executive order on October 30, 2023 about the safe development of AI. However, this is not the first AI-related guidance this administration has published. They previously issued a Blueprint for an AI Bill of Rights and an AI Risk Management Framework. But both the blueprint and framework suffer from the same fatal flaw: they were voluntary recommendations to follow, and thus unenforceable. The question is then, is this executive order any better?

The executive order begins by promising that the federal government will take many actions to protect against AI. One of the first promises is that the federal government will create a labeling mechanism so people know when content is generated. While that seems nice in theory, there is no practical way for the government to fulfill this promise. Many AI-generated content detectors already exist, such as GPTZero, Undetectable AI, and Sapling, however none of them are 100% accurate. I even tested this by putting this article in multiple detectors, and half of them detected AI content. A recent study by the University of Maryland also mathematically proved that as AI improves, it will become impossible to differentiate with certainty between AI-generated and human created content. Furthermore, AI-generated content detectors face issues of bias against non-native English speakers.

Some of the other promises include protecting data privacy, combating discriminatory use of AI, and investing in AI-related education and job training. While these goals are important, they are all promises to do something in the future. There is no definite plan or explanation on how the government will fulfill these promises. Instead, the order gives a variety of regulatory agencies and agency heads various deadlines to come up with a plan on how to fulfill these promises. And we have seen similar requirements on regulatory agencies from executive orders not being followed. For example, in 2019, Former President Trump signed an executive order about AI after which only the Department of Health and Human Services implemented a plan following the requirements of the executive order. So what makes President Biden’s executive order any more likely to be followed?

While the executive order makes a lot of promises, it also takes two important steps. The first is that all companies will be required to disclose the results of their safety tests on AI to the government. While some large companies had already voluntarily agreed to this, it is now a mandatory requirement for every company developing AI that could pose a risk to US health and safety. However, one issue with this requirement is that there is no guidance on what companies should do if their AI fails a safety test. If the AI is not open to the public to use, then it is clear that more development should be done. But can they nonetheless release it to the public? And if the AI is open to the public already, what does the company have to do? Do they have to take the AI down until it passes the safety tests? Do they have to put a disclaimer somewhere that the AI failed the safety tests? The lack of guidance is a recurring issue in this executive order.

The second important step is the establishment of a White House AI Council consisting of various Cabinet members and heads of regulatory agencies. A coordinated oversight body could be instrumental in ensuring that the federal government is fulfilling its promises. However, there is no information on what exactly this council will do. All the executive order says is that the council will coordinate the efforts of the different agencies across the federal government. But the efficacy of these efforts remains to be seen.

Going back to the original question, is this executive order any better? Only slightly. The executive order is weak. It makes a lot of promises without anything to back it up. And the steps it does take are not fully explained. There is no guarantee this will change anything. But it does start the process of regulating AI. So, what should the federal government do next?

 

The European Union AI Act

The EU has consistently been ahead of the US regarding technology regulations benefiting consumers. Last year, they passed regulation standardizing charging ports on all electronic devices, which is why Apple is once again changing their charging wire. And earlier this year, the European Parliament voted to regulate loot boxes in video games. Now, they are working on an exhaustive AI act that will regulate various kinds of AI.

The key part of this proposed regulation is dividing all AI into four different risk categories, based on the type and capabilities of each AI, and regulating them appropriately.

The first category is unacceptable risk. It would include AI with social scoring systems or systems that remotely monitor people in real time in public spaces. These AIs would have a blanket ban and be completely prohibited.

The second category is high risk. It would include AI falling under the EU’s product safety legislation, like medical devices, and AI used in eight specific fields, like biometric identification and law enforcement. These AIs would be heavily regulated, with obligations placed on both the creators and the users. Some requirements for AI in this category include establishing and maintaining a risk management system, ensuring human oversight, and using AI according to their proper instructions.

The third category is limited risk. It would include generative AI, chatbots, deepfakes, and some others. These AIs would have transparency obligations so users know they are interacting with AI or viewing deepfakes.

The fourth and final category is minimal or no risk. It includes all other AI, and they would be completely unregulated because of their low risk level.

Under this four-tier category system, the EU attempts to tailor regulations to match the capabilities and risks of different types of AI. Furthermore, once passed, this regulation would be enforceable against companies creating AI.

 

Is More Regulation the Right Answer?

Some critics may argue that regulating AI would stifle innovation and progress. But that argument neglects to recognize the important risks that AI poses. For one, AI is not trustworthy because it will make up facts. This has caused lawyers to accidentally use fake court cases when they asked AI a legal question. A second risk is that AI will expose confidential or personal information, especially when AI is trained using personal information from the web. Another risk is that AI will further bias and discrimination. This has already happened when Microsoft’s AI started spouting racist slurs. Without regulation, companies are free to create risky AI. But effective regulation would reduce these risks without stifling innovation, because you do not need to have an AI make up facts, expose personal information, and further bias to innovate.

 

Conclusion

As Will McAvoy, from HBO’s The Newsroom, once said, “The first step in solving any problem is recognizing there is one.” And President Biden’s executive order is clear recognition from the government that there is a problem: the US needs to start regulating AI. But it is only the first step. We need to follow the EU’s example. We need specific, enforceable legislation from Congress detailing different categories of AI and their requirements. Otherwise, the public will be left defenseless.

 

Suggested Citation: Akshat Shah, Is an Executive Order on AI Enough?, Cornell J.L. & Pub. Pol’y, The Issue Spotter (November 27, 2023), http://jlpp.org/blogzine/is-an-executive-order-on-ai-enough/.

 

Akshat Shah is a second-year law student at Cornell Law School. He graduated from Rensselaer Polytechnic Institute with degrees in computer science and economics. Aside from his involvement with Cornell Law School’s Journal of Public Policy, Akshat is the co-vice president of the South Asian Law Student Association.


Deprecated: file_exists(): Passing null to parameter #1 ($filename) of type string is deprecated in /home/r0bfc7luszh6/public_html/blogzine/wp-includes/comment-template.php on line 1616