Who we are

Soul Machines™ is a ground-breaking high tech company of AI researchers, neuroscientists, psychologists, artists and innovative thinkers; re-imagining how we connect with machines.

We bring technology to life by creating incredibly life-like, emotionally responsive artificial humans with personality and character that allow machines to talk to us literally face-to-face!

Our vision is to humanize artificial intelligence to better humanity.

what we do

We use neural networks that combine biologically inspired models of the human brain and key sensory networks to create a virtual central nervous system that we call our Human Computing Engine™.

When you 'plug' our engaging and interactive artificial humans into our cloud-based Human Computing Engine™, we can transform modern life for the better by revolutionizing the way AI, robots and machines interact with people.

When we, as human beings interact face-to-face, it's on the basis of both intellectual and emotional engagement. It's in our DNA. It's something we do naturally. What if machines were able to do this with us as well?

With their unprecedented level of intelligence and natural expressions, our life-like artificial humans can connect with us in a much more human way. By analyzing reactions and learning in real time they not only recognize emotional expressions but respond appropriately and interactively.

Our emotionally intelligent artificial humans are opening the doors to a new-era of human-style customer experience that can be utilized across a wide range of industries that are leading the way in this era of digital disruption - including Automotive, Financial Services & Banking, Healthcare, Media, Software and Technology.

Kevin Hanley - Director of Innovation, Royal Bank of Scotland Group - shares how CORA, an artificial human created for UK bank NatWest, will change the way customers can get quick, accurate answers on everyday banking questions.

Soul Machines™ has entered into a new partnership with Daimler Financial Services to create Sarah, an artificial human designed to help customers with personalized assistance for the company's services, including car financing, leasing and insurance.

The 3D Faces we create are as close to the real thing as we can make them. They are the most important instrument of emotional expression and engagement between people. We model the face in detail from the way the facial muscles create complex expressions all the way through the eyes that reflect what they see. We are developing full bodies from our artificial humans with the same physiological control systems. Our artificial humans are perfect for AR and VR.

Personality. Every one of our artificial humans comes with its own personality. We create the character behind the face based entirely on the role the artificial human will have in the "real" world. If for example, the artificial human will be a virtual customer agent we will incorporate a range of emotional responses, expressions, and behaviors that are consistent with the role and the core values of the organization that they will be representing.

Neural

Network

Models

Our neural network models are at the core of our embodied cognition platform. They were originally developed in our BabyX research program to give BabyX the ability to learn in real-time, express, speak and then recognize words and objects. Processing information from sensory inputs to generate behavior, these neural system models enable our artificial humans to express themselves based on the people they interact with.

We have developed biologically inspired models of the brain that are responsible for some of the key capabilities of our artificial humans. These are controllable by virtual neurotransmitters and hormones like dopamine, serotonin, and oxytocin. Together they influence virtual physiological states which guide learning and behavior, modulating the emotions that our artificial humans "feel" and express.

Intelligent sensors provide our artificial humans with the ability to see via a webcam and hear via the microphone in the device are just the beginning of the virtual nervous system that controls many of the physiological inputs that help bring our artificial humans to life.

Visual &

Auditory

Systems

Visual and auditory systems provide the data that feeds our identification, emotion detection, and analysis systems. Our auditory systems are also responsible for providing the captured voice stream to the natural language processing (NLP) engine which in turn asks questions of the AI platform.

Voice and speech are created specifically for an artificial human depending on the language and or accent required. To ensure the most life-like facial expressions while talking we train the muscles and lip movement to match the voice. We have even provided people with deafness the ability to lip read.