Real Time Projection Keystoning Based on Location of Viewer

I'm trying to find a way to correct the perspective of a projection which shows a room looking through the vantage point of a window. The viewers will be appx. 20' away from the screen and will be moving left to right. I would like to have the image account for the perspective change as they move left to right. Is there a method to capture the location of the viewers through a camera and drive a skew on the x axis based on their position in the room?

I think this was done in the most recent version of Mission Impossible ( the movie series).
There they used an eye trackingsystem, which is good only for a single viewer.
As everyone will be in a different position their perspective will change.

In reality I don't think this has actually reached actual tech yet, or not at a achievable price point.

I agree with @Amiers. This would be the best method for all possible viewers. This is the basic principle of the eyes that follow you in paintings/sculptures.

The opinions expressed by “ruinexplorer” do not in any way reflect those of his employers or any related partnerships. His given name is withheld to help to prevent his identity being established with his employer, so that he is not misconstrued as being an official voice for them.

If you really want a tracking thing for some art installation. I would investigate XBox one connect. I have no useful info but I am sure there are people playing with them doing this type of capture. Then you could use a live video rendering software like isadora to take some imput out of connect into the x values. I would look into communities playing with connect to see what hacks they have put together.

also Wii motes could do things... if you can put IR LEDs on the viewers

MAX/MSP can do it. It takes a semi steep learning curve, but you can use almost any type of data to effect, generate or report other data. I once did a project where I mapped a standard laptop camera to specific areas of the body making each area an instrument. Arms were a piano, chest was a kick drum, legs were guitars. Depending on the velocity and distance of movement the pitch and frequency were effected. Similarly you could map "zones" of your image to effect the x axis in correlation to the image. I did the project as part of an independent study in Electronic Music Engineering in grad school, it's a very amazing piece of software, but it's not for the faint of programming.