U.S Vice President, Kamala Harris is slated to meet on Thursday, May 4, 2023, with the Chief Executive Officers of four major companies developing artificial intelligence as the Biden administration rolls out a set of initiatives meant to ensure the rapidly evolving technology improves lives without putting people’s rights and safety at risk.
The Democratic administration plans to announce an investment of $140 million to establish seven new AI research institutes, administration officials told reporters.
Moreover, the White House Office of Management and Budget is expected to issue guidance in the next few months on how federal agencies can use AI tools.
There will also be an independent commitment by top AI developers to participate in a public evaluation of their systems in August at the Las Vegas hacker convention, DEF CON.
Harris and administration officials plan to discuss the risks they see in current AI development with the CEOs of Alphabet, Anthropic, Microsoft and OpenAI. The government leaders’ message to the companies is that they have a role to play in reducing the risks and that they can work together with the government.
Authorities in the United Kingdom also are looking at the risks associated with AI. Britain’s competition watchdog disclosed that it is opening a review of the AI market, focusing on the technology underpinning chatbots like ChatGPT, which was developed by OpenAI.
Last month, President Joe Biden noted that AI can help to address disease and climate change but also could harm national security and disrupt the economy in destabilizing ways.
Ethical And Societal Concerns About AI
The release of the ChatGPT chatbot this year has led to increased debate about AI and the government’s role with the technology. Due to the fact that AI can generate human-like writing and fake images, there are ethical and societal concerns.
OpenAI has been secretive about the data its AI systems have been trained upon. That makes it hard for those outside the company to understand why its ChatGPT is producing biased or false answers to requests or to address concerns about whether it’s stealing from copyrighted works.
Margaret Mitchell, Chief Ethics Scientist at AI startup Hugging Face, noted in an interview that companies that are worried about being liable for something in their training data might also not have incentives to properly track it.
“I think it might not be possible for OpenAI to actually detail all of its training data at a level of detail that would be really useful in terms of some of the concerns around consent and privacy and licensing. From what I know of tech culture, that just isn’t done.”
Margaret Mitchell
“I think it’s really going to be up to the governments to decide whether this means that you have to trash all the work you’ve done or not,” Mitchell opined.
“Of course, I kind of imagine that at least in the U.S., the decisions will lean towards the corporations and be supportive of the fact that it’s already been done. It would have such massive ramifications if all these companies had to essentially trash all of this work and start over.”
Margaret Mitchell
Theoretically, at least, some kind of disclosure law could force AI providers to open up their systems to more third-party scrutiny. However, with AI systems being built atop previous models, it will not be easy for companies to provide greater transparency after the fact.