“Imagine if we had a system that could keep account of all the objects that we interact with in our daily lives,” the researchers said.

“By keeping track of the locations of the objects, we could build a smart search engine for our home that could answer queries like – where are my eye glasses, or my TV-remote, or my wallet?”

Although alternative solutions, such as the use of radio-frequency identification chips already exists, the men said their system was many times cheaper due to the high cost of RFID readers.

What goes where

The researchers noted that running a computer program that simultaneously tracked all the owner’s objects in real-time would be too processor-intensive.

So they based their design around the principle that objects only change locations when humans move them.

As a result the Kinect sensor system focuses on tracking human figures and then looking for objects that have changed position in their vicinity.

Although the Kinect sensor’s capabilities are limited – it only sees objects up to 11 feet (3.4m) away and only provides “skeleton data” at 15 frames/second – the Kinsight program has commonsense notions built into it to improve accuracy: so it knows that a coffee cup is most likely to be found at a study desk, or kitchen sink, but not inside a bath.

“This means that, when in doubt, an object recognition algorithm can use this knowledge to identify an object by analysing the likelihood of it being at some location, or looking for the candidate objects in their other locations,” the researchers said.

On the move

Algorithms for Kinect sensor were also created to help the computer learn the appearance of objects and the context they were likely to be used in by analysing the data gathered.

To prove the Kinect sensor system would worked, the two scientists labelled 48 objects – including knives, forks, keys and a Rubik’s cube – and identified 80 possible locations around a house.

They then asked volunteers to move the items around according to randomly generated patterns.

The results suggested room for improvement – errors were more likely if the objects were very small, far away, transparent or placed too closely together – but the team said these problems should be addressed by using more sensors per room and adopting more sensitive depth-cameras.

In the meantime, they say that even when the program does lose track of possessions, it can still say were they were last seen which may still prove helpful.

NOTE: This article was first published on BBC News. Click HERE to view the original article.

POPULAR CATEGORY

Welcome to eVolved by WorldNgayon, an online news portal that features stories about development, innovation and change. WorldNgayon comes from two words “World” and “Ngayon” (a Filipino word meaning now).