Do You Believe Your Artificial Intelligence

Do You Believe Your Artificial Intelligence?

One of the most transformation technologies of our era is artificial intelligence (AI). Security risks and pertinent questions should be foreseen and elucidated before the tech starts to conquer the world. That’s why AI security-related consideration is of utmost importance.

After a new invention emerges, people become worried about safety issues — sometimes too late. Have a look at the advancement of different technologies, starting with networks, which began to grow from the 1980s. The increase of network security solutions proceeded in the 1990s. Afterward, an age of personal computers and anti-viruses began in the 2000s. Afterward, there were applications and the program safety boom in the 2010s. In terms of AI, since the second half of this 2010s was the age of smart technology, in the 2020s, we will likely find a requirement for AI safety.

The matter is, we should care about this possible future today. A number of critical incidents have taken place, like the Tesla car incident, and thousands of research papers have been released. I recently published several technical articles describing why AI safety is presently important, but let’s look at distinct steps that have been taken from the initiatives and regulations area.

International AI regulations

There’s a growing international effort to develop regulations and ethical guidelines for AI.

These plans feature some countries’ approaches to AI and ethics, but also mention security as part of those initiatives.

Graduating Throughout A Pandemic

The U.S. is one of the first countries to talk about the possible problems of AI safety, such as adversarial attacks, in its own AI plan”The National Artificial Intelligence Research And Development Strategic Plan: 2016.” One of the strategies in the document was to guarantee the security and safety of AI systems at each stage of the AI system life cycle. Later, many leading nations published similar documents where safety was one of the contemplated issues.

In 2017, the U.K. government interim strategy believed AI to be a key technology fad and an essential tool in identifying and responding to security risks. As an example, the authorities published a group of high-tech security principles for automatic and connected vehicles (CAV) and intelligent transport systems (ITS) and smart cities, in which it identified what good cybersecurity resembles. In addition, it developed an automotive-specific framework for a security assessment to assist the industry benchmark its products during the design and development stage and worked out a guide on how to manage risks in a distribution chain.

Canada’s focus on integrity quickly led to some of the earliest international AI ethical principles” Pan-Canadian Artificial Intelligence Strategy” was printed on March 22, 2017, under the direction of the Canadian Institute for Advanced Research (CIFAR).

On March 31, 2017, in its”Artificial Intelligence Technology Strategy,” Japan-focused mainly on the cultural and social aspects of artificial intelligence development, and security was cited among the four priority areas together with health, productivity, and mobility.

In April 2017, China generated many state-supported AI governance and integrity initiatives.

Since that time, more than 15 countries, such as Singapore, South Korea, UAE, and others, have published various documents mentioning AI security, privacy, security, and trustworthiness.

AI security initiatives

A lot of AI safety initiatives have been launched in the first quarter of 2019 from the U.S. and the European Union:

These records consider theories to create robust machine learning versions resistant to attacks as well as examine the landscape of feasible AI threats and ways to stop or better mitigate those dangers. Various regions of policy and governance are applied to AI and its implementations, including technology and data protection.

In 2020 and 2021, the worldwide dissemination of all AI security records is anticipated. Further research could be conducted to analyze emerging AI security solutions and initiatives. Along with being aware of AI-security initiatives, we will need to follow recommendations for implementation and train AI programmers, cybersecurity specialists and IT teams about the best way best to operationalize AI security in associations or take more practical steps like AI security examinations.

As seen above, the amount of initiatives is increasing and is expected to grow even more in the future. Taking into consideration that early adopters started to release their own strategies in 2017 and were prepared to show detailed documents on the trustworthiness and safety of AI just a couple of decades later in 2019, we can predict that many of the countries that joined later with their AI initiatives will behave in precisely the exact same way.

We expect that those who published AI strategies in 2018 will show their own documents on the subject of AI security, security, and hope from 2020 and 2021. Also, some of these concepts will be able to form the basis of authentic law demands in the not too distant future, not only vague ideas on the topic.

Post a Comment