Home News Playing with fire: How to adapt to the new realities of AI

Playing with fire: How to adapt to the new realities of AI

by WeeklyAINews
0 comment

Be a part of prime executives in San Francisco on July 11-12, to listen to how leaders are integrating and optimizing AI investments for fulfillment. Learn More


When people found fireplace roughly 1.5 million years in the past, they most likely knew they’d one thing good immediately. However they seemingly found the downsides fairly rapidly: Getting too shut and getting burned, by accident beginning a wildfire, smoke inhalation and even burning down the village. These weren’t minor dangers, however there was no going again. Fortuitously, we managed to harness the facility of fireside for good.

Quick forwarding to right now, synthetic intelligence (AI) may show to be as transformational as fireplace. Like fireplace, the dangers are enormous — some would say existential. However, prefer it or not, there isn’t a going again and even slowing down, given the state of world geopolitics.

On this article, we discover how we are able to handle the dangers of AI and the completely different paths we are able to take. AI isn’t just one other technological innovation, it’s a disruptive pressure that can change the world in methods we can’t even start to think about. Nevertheless, we have to be conscious of the dangers related to this know-how and handle them appropriately.

Setting requirements for using AI

Step one in managing the dangers related to AI is setting requirements for using AI. This may be finished by governments or business teams, and they are often both obligatory or voluntary. Whereas voluntary requirements are good, the truth is that the businesses which might be probably the most accountable are likely to observe guidelines and steerage, whereas others pay no heed. For overarching societal profit, everybody must observe the steerage. Subsequently, we suggest that the requirements be required, even when the preliminary commonplace is decrease (that’s, simpler to fulfill).

See also  Amazon finds itself in the unusual position of playing catch-up in AI

As as to whether governments of business teams ought to prepared the ground, the reply is each. The truth is that solely governments have the heft to make the principles binding, and to incentivize or cajole different governments globally to take part. However, governments are notoriously slow-moving and liable to political cross-currents — undoubtedly not good in these circumstances. Subsequently, I consider that business teams have to be engaged and play a number one function in shaping the pondering and constructing for the broadest base of help. Ultimately, we want a public-private partnership to attain our targets.

Governance of AI creation and use

There are two issues that have to be ruled with regards to AI: Its use and its creation. The usage of AI, like all technological improvements, can be utilized with good intentions or with unhealthy intentions. The intentions are what issues, and the extent of governance ought to coincide with the extent of threat (or whether or not inherently good, or unhealthy, or someplace in between). Nevertheless, some varieties of AI are inherently so harmful that they have to be fastidiously managed, restricted or restricted.

The truth is that we don’t know sufficient right now to put in writing all of the laws and guidelines, so what we want is an efficient start line and a few authoritative our bodies that can be trusted to subject new guidelines as they turn out to be vital. AI threat administration and authoritative steerage have to be fast and nimble; in any other case, it would fall far behind the trail of innovation and be nugatory. Current industries and authorities our bodies transfer too slowly, so new approaches have to be established that may proceed extra rapidly.

See also  Google upgrades Bard to compete with ChatGPT

Nationwide or international governance of AI

Governance and guidelines are solely nearly as good because the weakest hyperlink. The buy-in of all events is important. This would be the hardest facet. We should always not delay something to attend for a worldwide consensus, however on the identical time, international working teams and frameworks needs to be explored.

The excellent news is that we aren’t ranging from scratch. Numerous international teams have been actively setting forth their views and publishing their output; notable examples embody the not too long ago launched AI Risk Management Framework from the U.S.-based Nationwide Institute for Science and Know-how (NIST) and Europe’s proposed EU AI Act — and there are lots of others. Most are of a voluntary nature, however a rising quantity have the pressure of regulation behind them. In my opinion, whereas nothing but covers the complete scope comprehensively, for those who have been to place all of them collectively, you’ll be at a commendable start line for this journey.

Reflecting

The experience will certainly be bumpy, however I consider that people will finally prevail. In one other 1.5 million years, our ancestors will look again and muse that it was robust, however that we finally obtained it proper. So let’s transfer ahead with AI, however be conscious of the dangers related to this know-how. We should harness AI for good, and take care we don’t burn down the world.

Brad Fisher is CEO of Lumenova AI.

Source link

You may also like

logo

Welcome to our weekly AI News site, where we bring you the latest updates on artificial intelligence and its never-ending quest to take over the world! Yes, you heard it right – we’re not here to sugarcoat anything. Our tagline says it all: “because robots are taking over the world.”

Subscribe

Subscribe my Newsletter for new blog posts, tips & new photos. Let's stay updated!

© 2023 – All Right Reserved.