Nice Shoes, Wanna Recognize Some Input?

Even though giant multouch display tables have been around for a few years now we have yet to see them being used in the wild. While the barrier to entry for a Microsoft Surface is very high, one of the biggest problems in implementing a touch table is one of interaction; how exactly should the display interpret multiple commands from multiple users? [Stephan], [Christian], and [Patrick] came up with an interesting solution to sorting out who is touching where by having a computer look at shoes.

The system uses a Kinect mounted on the edge of a table to extract users from the depth images. From there, interaction on the display can be pinned to a specific user based on hand and arm orientation. As an added bonus the computer can also recognize users from their shoes. If a user is wearing a pair of shoes the computer recognizes, they’ll just walk up to the table and the software will recognize them.

Continue reading “Nice Shoes, Wanna Recognize Some Input?”

Augmented Reality Ex Nihilo

[David] sent in a nice project to demonstrate augmented reality with ARtoolkit and discuss the deep philosophical underpinnings of the meaning of nothingness. The good news is he was able to create a volume control button on a sheet of paper with a marker. The bad news is the philosophical treatment is a bit weak; [David] built something cool, so we’re able to let that slide for now.

This build was inspired by the Impromptu Sound Board made using a Kinect and a piece of paper. The idea behind the sound board is simple – draw some buttons on the paper, and use them to play short sound clips. [David] took this idea to make a small tutorial on augmented reality for Occam’s Razor.

The hardware is very simple – just a webcam, a piece of paper, and a marker. After [David] draws a large square on the paper, the code recognizes it as a volume control. Rotating the paper counterclockwise increases the volume, and clockwise turns the volume down. It’s a neat build to get into the foundations of augmented reality.

Check out the video demo of [David]’s build after the break.

Continue reading “Augmented Reality Ex Nihilo”

Web-enabled Kinect

There are Kinect hacks out there for robot vision, 3D scanners, and even pseudo-LIDAR setups. Until now, one limiting factor to these builds is the requirement for a full-blown computer on the device to deal with the depth maps and do all the necessary processing and computation. This doesn’t seem like much of a problem since [wizgrav] published Intrael, an HTTP interface for the Kinect.

[Eleftherios] caught up to [wizgrav] at his local hackerspace where he did a short tutorial on Intrael. [wizgrav]’s project provides each frame from the Kinect over HTTP wrapped up in JSON arrays. Everything a Kinect outputs aside from sound is now easily available over the Internet.

The project is meant to put computer vision outside the realm of desktops and robotic laptops and into the web. [wizgrav] has a few ideas on what his project can be used for, such as smart security cameras and all kinds of interactive surfaces.

After the break, check out the Intrael primer [wizgrav] demonstrated (it’s Greek to us, but there are subtitles), and a few demos of what Intrael ‘sees.’

Continue reading “Web-enabled Kinect”

diy-laser-scanner

Cheap DIY Laser Scanner Is Quite Impressive

With the introduction of the Kinect, obtaining a 3D representation of a room or object became a much easier task than it had been in the past. If you lack the necessary cash for one however, you have to get creative. Both the techniques and technologies behind 3D scanning are somewhat complicated, though certainly still within reach as maker [Shikai Chen] shows us. (Google Translation)

He wanted to create 3D scanned images, but he didn’t have the resources to purchase a Kinect. Instead, he built his own scanner for about 1/6th the cost. Interestingly enough, the scanner resembles what you might imagine a very early Kinect prototype looked like, though it functions just a little bit differently than Microsoft’s creation. The scanner lacks any sort of IR emitter/camera combo, opting to use a laser and a USB VGA camera instead. While scanning, the laser shines across the target surface, and the reflected light is then picked up by the camera.

So how does this $25 DIY laser scanner measure up? Great, to be honest. Check out the video below to see how well his scanner works, and be sure to take a look through his second writeup (Google Translation) as well for more details on the project.

[via Seeedstudio]

[youtube=http://www.youtube.com/watch?v=jLZ-s9KRzG8&feature=player_embedded&w=470]

Think You Can Take Kinect To The Next Level? Check Out Kinect Accelerator

kinect-accelerator

If you’ve got a crazy ingenious idea for Microsoft’s Kinect peripheral, but don’t have the means to make your dream a reality, the Kinect Accelerator just might be the opportunity you’ve been waiting for.

Microsoft, having performed a complete 180-degree turnaround from their initial stance on Kinect hacking, is embracing developers more than ever with this new program. They are offering a $20,000 along with development space to ten startup companies, in hopes of turning out some incredible Kinect applications. At the end of the three month program, each group will have the opportunity to present their creations to a group of angel investors, which is a fantastic opportunity.

Obviously competition to gain entry into the program will be pretty fierce, but if you think you have what it takes, get your application in now. Judging by the Kinect Accelerator FAQ section, this looks to be something geared towards small tech startups rather than individuals, but it never hurts to give it a shot.

A Kinect Primer

Yes, the Kinect is over one year old now, and after some initial unhappiness from [Microsoft], it’s become a hacker’s best friend. [Eric] decided to celebrate this with an Article all about how it works.  If you’re new to this piece of hardware and want to get into working with it, this should be a good hacking introduction.  If you’ve been reading [HAD] lately, you will have noticed this information being used to “build a Kinect bot for 500 bones.”

Some interesting facts in this article include that the Kinect measures 307200 distance point, known as a “point cloud” in the gaming area. From this, it’s able to construct a 3D image of the environment around it and allow interaction. Such interesting hardware didn’t take long to hack after Adafruit announced a $3000.00 bounty to open it up to the masses.  This only took four days to do, making one wonder why, with their incredible resources, [Microsoft] wouldn’t either more effectively lock it down or officially open it to be hacked and modified to begin with. Our vote would be to officially open it up, but no one consulted us on the decision.

Build A Kinect Bot For 500 Bones

[Eric] sent in his tutorial on building a Kinect based robot for $500, a low-cost solution to a wife that thinks her husband spends too much on robots.

For the base of his build, [Eric] used an iRobot Create, a derivative of the Roomba that is built exclusive for some hardware hackery. For command and control of the robot, an EEE netbook takes data from the Kinect and sends it to the iRobot over a serial connection.

The build itself is remarkably simple: two pieces of angle aluminum were attached to the iRobot, and a plastic milk crate was installed with zip ties. The Kinect sits on top of the plastic crate and the netbook comfortably fits inside.

A few weeks ago, [Eric] posted a summary of the history and open-source software for the Kinect that covers the development of the Libfreenect driver. [Eric] used this same driver for his robot. Currently, the robot is configured for two modes. The first mode has the robot travel to the furthest point from itself. The second mode instructs the robot to follow the closest thing to itself – walk in front of the robot and it becomes an ankle biter.

There is a limitation of the Kinect that [Eric] is trying to work around. Objects closer than 19 inches to the Kinect appear to be very far away. This caused a lot of wall bumping, but he plans on adding a few ultrasonic sensors to fill the gap in the sensor data. Not bad for a very inexpensive autonomous robot.