Breaking News

When Artificial Intelligence Explains, Humans Learn – RTInsights

If artificial intelligence machines can take us through their experiences, they can begin to teach us new ways to solve problems.

We’ve spent so much time building machines that think the way we do. Now that we’ve partially accomplished that, it could be time to learn from our machines in ways we didn’t think possible. At the heart of this concept is to leverage the fact that many artificial intelligence applications learn over time as more data becomes available and outcomes are evaluated. If the AI systems could then share this gained knowledge with humans, computers could soon be responsible for our greatest innovative leaps.
Essentially, AI would
explain how and why it made a decision or took an action, and humans would
learn from this knowledge base. It is the equivalent of a new employee being
mentored by a seasoned professional.
See also: Decision-Making Algorithm Aids Group Choices
Artificial intelligence black boxes do us no favors
Traditionally, such a
process does not happen. AI is treated as a black box, revealing little about
the way machines come to decisions. We get amazing insights from millions and
billions of data points, but we cannot figure out how machines reach those
That’s a problem when we
need to understand new disease recommendations or figure out how machines
choose certain candidates over others. Researchers are beginning to run after
explainable AI, not just for liability or privacy purposes, but for the
learning opportunity it presents for humanity.
Machines can teach us from
Experience is a
wonderful teacher, but humans cannot hope to experience all the data needed to
reach certain conclusions. This is where reinforcement learning can fill the
Machines use
reinforcement learning to explore the world and come to conclusions based on
those experiences. If machines can take us through their experiences, they can
begin to teach us new ways to solve problems, new components of existing
problems, and a whole host of other things.
Currently, machines
simply spit out conclusions. We cannot retrace the steps. We cannot peer into
the process. In the early days, this wasn’t an issue; we were so enamored that
machines were thinking that it didn’t
matter how.
The pursuit of explainable AI
The pursuit of explainable AI means access to a lot of things. We would no longer have to scrap an entire program because it came to sketchy conclusions. We’d no longer be liable for a machine’s terrible decisions based on mysterious problems with data.
Even more than that, we
may finally make a huge jump in innovation. When machines can explain their
fantastic solutions based on data patterns beyond our control, we may find
ourselves on the cusp of a huge leap.