AI: Can A Machine Ever Be Human, Convincingly? ﻿

The inclusion of ‘learning abilities’ – mostly thought unique to humans and very few other evolved primates – defines artificial intelligence to a large extent. Faced with unfamiliar situations, how the program deals with the problems and attempts to solve them is key to identifying a stretch of software code as ‘artificially intelligent’.

Artificial Intelligence has made
the leap from science fiction to real life in a short matter of time. It was initially
envisioned as a panacea for the intricate but repetitive processes that aided scientific
research and technological advancement – a role it has fulfilled and, in many
instances, surpassed.

Training a program by making it
understand a variety of sensory inputs, whether in the form of digital or
analog data, does not mean that program has ‘intelligence’. The result of this
factor being used to decide the intelligence of software leads to various
technologies that were quite revolutionary at their inception now being
classified as routine programs, because their previously groundbreaking tasks
have become rudimentary in today’s advanced day and age.

A Brief History of AI

Automation has been a pursuit of humanity since classical Greek antiquity. The word ‘automaton’ itself is used by Homer to refer to machines acting according to their own will. There is ample evidence in literature and history that shows how we have striven to recreate machines that not only look like us, but walk, talk and act like us. The more successful efforts towards such aims are said to be in the ‘uncanny valley’, an uncomfortable state which results from the almost, but not entirely, accurate depiction of human beings by doppelganger machines.

Alan Turing was instrumental in making
artificial intelligence a practical field. Approaching AI in purely
mathematical binary terms, digitization was used as the platform to erect
expert systems, which use inference engines and knowledge bases to make
decisions. Moore’s Law, which predicted computing power rising up while
component sizes reduced, still remains applicable, albeit to a slightly lesser
extent.

Now, with data surging forth from
all sorts of sources right from our handheld devices to astronomical
observations and literal rocket science, machines that have been developed
specifically to ‘think like a human’ are rapidly being deployed in a variety of
fields, form bioengineering to synthetic medicine. Nearer our daily lives,
search engines [one (followed by a hundred zeros) in particular, but all of
them in general] and flagship smartphones use all the learnings gleaned from AI
to deliver ‘personalized experiences’ right into our hands!

We Are Already AI-ed, Daily!

In 2014, Stephen Hawking gave a
subliminal quote on AI: “It [AI] would
take off on its own and redesign itself at an ever increasing rate. Humans, who
are limited by slow biological evolution, couldn’t compete and would be
superseded.”

While such a day still seems far
off as of now, the quest for replicating human thought patterns and response
heuristics continues unabated. Programmers in diverse fields toil away every
day at their projects, attempting to reproduce the thought processes that make
up the human mind. They have to take many factors into consideration, not the
least of which is the ethical complication in ‘fooling’ a human into thinking
they are conversing – or, at basic levels, interacting – with a machine.

We are already carrying out a
great deal of everyday interactions with artificial intelligence. The level to
which it affects the technology in the palm of our hands is difficult to
identify at the user level. To delve deeper, we have to break down the integral
components of interactions amongst humans and machines – a task easier said
than done.

The question I asked at the
beginning is hard to answer, because it is rooted in the future. At Cyfuture,
we are accustomed to asking questions that require a certain kind of ‘never
giving in’ mindset to answer – for laterally solving problems or creating
innovative solutions to increase the effectiveness of existing legacy systems,
as well as drive businesses better.