Introduction to the Drone Journey

This post is the eleventh, and final, post in documenting the steps I went through on my journey to build an autonomous, voice-controlled, face recognizing drone. There are 10 other posts building up to this one which you can find at the end of this post.

Focus of this post

In this post I will share a video of the complete end-to-end demo and share details of the architecture which sits behind it. I will also share information on what I bought/used to bring this all together and relist all the different software, services and node packages in a single place.

Pulling It All Together

A lot of what we have been doing with this project is humanizing the way we communicate with machines/computers/things. That means talking and observing to drive intelligent interaction rather than using a mouse, keyboard or touch screen.

Our Autonomous Voice-Controlled, Face Recognizing, Drone is a smart drone which showcases, albeit crudely, how interaction with services filled with intelligence is going to evolve. It highlights the importance of cognitive services to the success of organizations in the future.

So with that said take a look at the entire end to end demo in the video below.

Introduction to the Drone Journey

This post is the tenth post in documenting the steps I went through on my journey to build an autonomous, voice-controlled, face recognizing drone. There are 9 other posts building up to this one which you can find at the end of this post.

Focus of this post

Up until now we have been mostly working on controlling the drone, using the Microsoft Cognitive Services Face API to identify people and lastly making use of the Microsoft Cognitive Services API to convert text to speech and speech to text.

Ultimately we will have built an intelligent end-to-end IoT solution featuring analytics and visualization. The instructions here can be used to also understand how to get data in from other devices as well!