Saturday, 28 October 2023

Majority Report: How Do We Monitor AI?

NAB

It is urgent that we regulate synthetic media and deepfakes before they undermine our faith in the truth, says Russell Wald, the director of policy for Stanford’s Institute for Human-Centered Artificial Intelligence.

article here 

“I’m concerned about synthetic media, because of what will ultimately happen to society if no one has any confidence in what the veracity of what they’re seeing,” he says in an interview with Eliza Strickland at IEEE Spectrum about creating regulations that are able to cope with the rapidly evolving technology.

“You’re not going to be able to necessarily stop the creation of a lot of synthetic media but at a minimum, you can stop the amplification of it or at least put on some level of disclosure, that there is something that signals that it may not be in reality what it says it is,” he says.

The other area that Wald thinks would help in terms of overall regulation is greater transparency regarding foundation data models.

“There’s just so much data that’s been hoovered up into these models, [but] what’s going into them? What’s the architecture of the compute? Because at least if you are seeing harms come out at the back end, by having a degree of transparency, you’re going to be able to [identify the cause].”

Of calls for regulation coming from AI developers themselves, Wald is scathing, “For them, it really comes down to would they rather work now to be able to create some of those regulations versus avoiding reactive regulation. It’s an easier pill to swallow if they can try to shape this at this point.”

What he would really like to see is great diversity of viewpoint in the discussions and decision-making process, not just from those in the tech industry, but from academics like himself and from law makers.

“Others need to have a seat at the table. Academia, civil society, people who are really taking the time to study what is the most effective regulation that still will hold industry’s feet to the fire but allow them to innovate?

This would mitigate the risk of inherent bias in certain algorithms on which decisions in judicial systems or legal systems or medical contexts might be based.

Like many academics with knowledge of the subject, Wald calls for a balanced approach. AI does have significant upside for humans as a species he says, pointing out the unprecedented ability of AI to sift through and test data to find solutions for diseases.

“At the same time, there’s the negative that I am truly concerned about in terms of existential risk. And that is where the human comes into play with this technology. Synthetic biology, for instance, could create agents that we cannot control. And there can be a lab leak or something that could be really terrible.”

Having given a precis of what is wrong, Wald turns to potential solutions by which we might regulate our way out of potential disaster. This is multi-pronged.

“First, I think we need more of a national strategy, part of which is ensuring that we have policymakers as informed as possible. I spend a lot of time in briefings with policymakers and you can tell the interest is growing, but we need more formalized ways of making sure that they understand all of the nuances here,” he says.

“The second part is we need infrastructure. We absolutely need a degree of infrastructure that ensures we have a wider degree of people at the table. The third part of this is talent. We’ve got to recruit talent and that means we need to really look at STEM immigration, and see what we can do because at least within the US the path for those students who can’t stay here, the visa hurdles are just too terrible. They pick up and go, for example, to Canada. We need to expand programs like the intergovernmental personnel act that can allow people who are in academia or other nonprofit research to go in and out of government and inform governments so that they’re more clear on this.”

The final piece in Wald’s argument is to adopt regulation in a systematic way. For this, he looks to the European Union, which is one of the most advanced territories in terms of formulating an AI Act. However, this is not expected to be ratified for at least another year.

“Sometimes I think that Europe can be that good side of our conscience side and force the rest of the world to think about these things. This is Brussels effect — which is the concept Europe has such a large market share, that they’re able to force through their rules and regulations, being among the most stringent and it becomes the model for the rest of the world.”

He identifies the UK’s approach to AI regulation as a potential model to follow because it seems to be more balanced in favor of innovation.

“The Brits have a proposal for an exascale computing system [to] double down on the innovation side and, where possible, do a regulatory side because they really want to see themselves as the leader. I think Europe might need to look into as much as possible, a degree of fostering an environment that will allow for that same level of innovation.”

Wald’s concern that AI will stem innovation is not to protect the larger companies, who can look after themselves, he says, but the smaller players might not be able to manage to continue if the law is too stringent.

“The general public should be aware that what we’re starting to see is the tip of the iceberg,” he warns. “There’s been a lot of things that have been in labs, and I think there’s going to be just a whole lot more coming.

“I think we need to have a neutral view of saying there are some unique benefits of AI for humanity but at the same time, there are some very serious dangers. So the question is how can police that process?”


No comments:

Post a Comment