[John B] is a software engineer and had some spare time on his hands, so he started messing around with his Kinect which had been sitting unused for awhile. He wanted to see what he could create if he was able to get Kinect data into a virtual environment that supported real-world physics. The first idea that popped into his head was to interface the Kinect with Garry’s Mod.
If you are not familiar with Garry’s Mod, it is a sandbox environment built on top of Valve’s Source engine. The environment supports real-world physics, but beyond that, it pretty much lets you do or build anything you want. [John] found that there was no good way to get Kinect data into the software, so he built his own.
He used OpenNI to gather skeletal coordinate data from Kinect, which was then passed to some custom code that packages those coordinates inside UDP packets. Those packets are then sent to a custom Lua script that is interpreted by Garry’s Mod.
The result is just plain awesome as you can see in the video below. Instead of simply playing some random game with the Kinect, you get to design the entire experience from the ground up. The project is still in its infancy, but it’s pretty certain that we’ll see some cool stuff in short order. All of the code is available on github, so give it a shot and share your videos with us.