Let Us Protect Ourselves From AI’s Dark Side!
What is Artificial Intelligence?
Now, Artificial Intelligence
or popularly known by its short-form AI, is a wide-ranging branch of computer
science that concerns itself with building smart machines that are capable of
performing tasks that typically require human intelligence. AI is also an
interdisciplinary science with multiple approaches, but the greater
advancements in machine learning and deep learning are creating a major impact
in almost every sector of the technology industry. Some common examples of AI
include automated customer
support, personalized shopping
experience, smart cars,
drones and much more.
Potential risks of AI
The promise of AI is real, but the potential risks also are simply as real. However, in the end appropriate steps taken will
reduce the risks. Artificial
intelligence and now augmented intelligence have received quite a lot
of attention. While some components may be overhyped, the era is a certain part of our future.
Adobe’s 2018 Digital Trends record found that while just 15% of corporations use AI today; 31% have it on their schedule for the following 365 days. Advancements in AI, such as machine learning and
neural networks, are marching us in
the direction of a greater interconnected and automated future. While the technology maintains to grow, the technology behind AI remains looking to apprehend how
the human mind works,
and replicate that
to assist the enhancement of our everyday life.
Considerations for an AI future
Despite the pleasant intentions for AI, computer structures may also in the end increase in ways we by no means imagined. This was illustrated at some point of the 2017 Neural Information Processing
Systems Convention whilst researchers provided an AI-based system for image mapping
that was found out to provide outcomes by using hiding source
data through Steganography.
The system produced
the outcomes researchers were looking for but with the aid of “cheating” and hiding data it needed to “succeed”. As AI continues to grow exponentially, the notion that this generation could doubtlessly be used for malicious purposes will become more actual. One latest example to bear
in mind are the consequences of
AI driving social
media bots at the 2016 US Election.
Ultimately AI will “research” primarily based on
the records we provide it to function on. We need to make sure we aren't biasing the effects with the aid of selecting information that fits our preconceived understanding of what is appropriate. Consider these instances where AI
misinterpreted data and came back with the wrong information.
The data security industry additionally has some work to do to create extra protection mechanisms
as advanced assaults on data integrity
are increasing and are equally hard to discover and shield against.
AI’s role in protecting the network
On the flip side, AI will hold to
play a role in protecting against different threats. There are corporations already developing AI-based products that provide threat-hunting, attack analysis and incident reaction to proactively look for potential
problems. Another instance is
the latest work of the New York Power
Authority to combine AI
into its energy grid.
While the machine can truly aid with detection of problems, it's critical that
the manner wherein the system responds to threats
is managed. As Rob
Lee said, “You don't want your grid operators, the humans
that are controlling the grid, to become so dependent on the machine learning
or AI model that they forget how to do their job."
Minimizing AI dangers
To be clear, AI may be used for excellent
purposes. However, it is critical for organizations to understand the whole image and to
figure out what can be performed to mitigate the risks.
Legislation: While studies and experiments should be unfettered to the greatest extent possible, before structures built on AI are given control of critical regions (e.G., critical infrastructure,
finance, healthcare, cyber protection, and so forth.) the results of failure must be described very well.
Bound the capabilities:With
the growing threats of AI, one way to control the risk of AI could be to set boundaries within
which the AI system must function to avoid the unnecessary consequences. If
limits are set by keeping in mind
the worst case situations,
controls may be installed that dictate exactly what the AI system can do. Examples of the worst-case scenarios would be stopping the automated delivery of medical treatment in a reaction to a detected
healthcare risk or proscribing the scale and frequency of economic transactions in response to strange market conditions.
Keep the human in
control of the ship:Similar to today’s research of self-driving vehicles, automated responses to network threats that
might be disruptive must be limited to a human’s decision, for now. Augmented
intelligence combines the velocity of device intelligence with the intuition and management of humans. Coupled with this is the want for the community to have the potential to evolve to external stimuli.
The AI machine can propose a course of action, or several. It could even expect the final results given a specific choice. But in the end the “red button” needs to be pushed with the aid of a human.
Experiment and install extra:
The first-class mitigation
for danger might also in truth be to boost up the studies and test in controlled regions. Using the structures in simulated or
replicated environments will allow researchers
to better understand while unintentional activities or
responses occur and also better understand how to mitigate them.
Training and deliberate layout:
Published frameworks on each social
and technical impacts and
consequences must be
made available to absolutely everyone. The promise of AI
is in fact real. However,
the potential risks also are simply as real.
As it is in the case
with countless other advances we have made,
countermeasures monitor themselves
and the benefits of
the new technologyultimately overtake
the danger. This too, could be the case with AI.
Conclusion
With the advancement in
technology, artificial intelligence might take over our world in no time.
However, certain measures and ethics must be followed to keep the relationship
between humans and artificial intelligence balanced.Just like every coin has
two sides, AI has both good and bad sides. On the dark side, there is a risk of
data security and data misinterpretation. But, with the necessary steps and
proper guidelines and regulations, the adverse effects of AI can be minimized.
Comments
Post a Comment