Recently Pivotal Labs decided to push mobile devices to the limit by connecting multiple devices to one other in order to have them behave as one. Through the use of image recognition, optical character recognition, persistent low cost connections and a whole lot of ingenuity Pivotal Labs was able to put together an innovative and unique experience. Below, we have collected all the accounts of the key individuals involved in bringing to life this piece of innovation and technology.

Second screen has been a buzzword for quite some time and rightfully so. Getting our tech gadgets to work as one has always been a desire. With the adoption of phones as the dominant personal computer over the last few years, we've naturally wanted to connect everything to them.

You've seen the reviews, the keynotes, the videos and the endless debates on whether this sci-fi tech-like innovation is a solution, a problem, or just an unfinished idea. Yet, we all can agree that Google's Project Glass, deemed to be the next hot personal computing device after iPhone's big break in 2007 and lauded to be at the forefront of technology for years to come, has the attention of the entire world. Consumers and professionals of all kinds are gawking at the ability to connect to the digital world with literally the blink of an eye.
Yet, Google Glass is still out of reach for most.

I have received many mocks for Android applications that were beautiful but did not translate appropriately to devices. The mocking/design process tends to happen separately or before the development process and reconciling the two can become a timely process. Designs also set expectations and imply flows that do not always translate to the development paradigms Android enforces. Here are a few concrete steps that will help you minimize some of the churn.

This page is a compliment to the Screen Size Android
application. The client code can be found on github as
well as the server code that receives and displays the
data. You can also use the following api call to pull
the full set of data down.

Computers are about to become a lot more human if IBM's annual "5 in 5" predictions are to be believed. The technology company's researchers have predicted that within five years computers will smell, hear and offer a touch experience that mimics real-world textures.

Some time has passed since my initial investigation of Google TV.
While this is still relevant, a lot is happening. OEMs are starting to
show their support with different Google TV implementations and Google has also kept their partners busy
with newer operating system releases.

With the traditional c model one could fork(). This would create a child and parent, the child would copy the parent's memory space and they would communicate through sockets, which were essentially files. vfork() would allow them to share the same memory space, but one at a time. Then there are also exec() and clone() that would similarly launch processes and allow for communication between them. Accurate communication proved to be very difficult, so eventually POSIX threads came to be. They brought mutex locks and semaphores to help with communication (locking). These are all very low end principles and implementing them takes a lot of care and consideration.

Google TV is a multimedia centre that takes in an HDMI feed and overlays it with the Android operating system. It attempts to connect to and communicate with a settop box (cable box) through the connecting HDMI cable. It reads channel information and controls the box by changing channels. Some Google TV devices, such as the Logitech Revue, have the ability to control your television as well. By integrating Google TV into a home theatre system, you have the ability to host native android applications and experience Chrome through your television. To interact with the device, users are given a keyboard and mouse in one controller. The controller is a standard keyboard which has had the arrow keys and numerical pad replaced with a digital arrow pad and a laptop-styled touch-pad mouse.

The idea behind droid physics was to offer a simple physics
environment that would aid early high school students to understand
some of the properties of physics. In particular, acceleration,
projectile motion and Friction.

Little Droid Creator was designed to showcase the dynamic engine
on which it runs. This document will first describe at a high level
what the game offers then we will dig down into the technical
aspects of the program and talk about how it offers the high level
aspects

Android is a framework that provides java programmers the ability
to control different aspects of smart devices. This interaction
happens through the Android SDK (Software Development Kit). Learn how to make an end to end experience with this tutorial.

StarFighter (JavaScript) available online for first time. This game was designed to show case the web development knowledge I had acquired while working with Visibli. One of the things to note is the async loop class. It demostrates how scope and threading works in JavaScript.