Pong In Real Life, Mechanical Pong

[Daniel Perdomo] and two of his friends have been working on a mechanical version of Pong for the past two years. We can safely say that the final result is beautiful. It’s quite ethereal to watch the pixe–cube move back and forth on the surface.

[Daniel] has worked in computer graphics for advertising for more than 20 years. However, he notes that neither he nor his friends had any experience in mechanics or electronics when they began. Thankfully, the internet (and, presumably, sites like Hackaday) provided them with the information needed.

The pong paddles and and pixel (ball?) sit onto of a glass surface. The moving parts are constrained to the mechanics with magnets. Underneath is a construction not unlike an Etch A Sketch for moving the ball while the paddles are just on a rail with a belt. The whole assembly is made from V-groove extrusion.

Our favorite part of the build is the scroll wheel for moving the paddle back and forth. For a nice smooth movement with some mass behind it, what’s better than a hard-drive platter? They printed out an encoder wheel pattern and glued it to the surface. The electronics are all hand-made. The brains appear to be some of the larger Arduinos. The 8-bit segments, rainbow LEDs, etc were build using strips glued in place with what looks like copper foil tape connecting buses. This is definitely a labor of love.

It really must be seen to be understood. The movement is smooth, and our brains almost want to remove a dimension when watching it. As for the next steps? They are hoping to spin it up into an arcade machine business, and are looking for people with money and experience to help them take it from a one-off prototype to a product. Video after the break.

Continue reading “Pong In Real Life, Mechanical Pong”

VR Telepresence Tank from Raspberry Pi, Google Cardboard, and Xbox Controller

It’s great to see different kinds of hardware and software tossed into a project together, allowing someone to mix things that don’t normally go together into something new. [Freddy Kilo] did just that with a project he calls his VR Robot Tank. It’s a telepresence device that uses a wireless Xbox controller to drive a tracked platform, which is itself headed by a Raspberry Pi.

The Pi has two cameras on a pan-tilt mount, and those cameras are both aimed and viewed via a Google Cardboard-like setup. A healthy dose of free software glues it together, allowing things like video streaming (with U4VL) and steering via the wireless controller (with xboxdrv). A bit of fiddling was required for some parts – viewing the stereoscopic cameras for example is done by opening and positioning two video windows just right so as to see them through the headset lenses. It doesn’t warp the image to account for the lens distortion in the headset, and the wireless range might be limited, but the end result seems to work well enough.

The tank is driven with the wireless controller while a mobile phone mounted in a headset lets the user see through the cameras; motion sensing in the phone moves those cameras whenever you move your head to look around. Remote Control hobbyists will recognize the project as doing essentially the same job as FPV setups for model aircraft (for example, Drone Racing or even Snow Sleds) but this project uses a completely different hardware and software toolchain. It demonstrates the benefits of having access to open tools to use as virtual “duct tape”, letting people stick different things together to test a concept. It proves almost anything can be made to work if you have a willingness to fiddle!

Continue reading “VR Telepresence Tank from Raspberry Pi, Google Cardboard, and Xbox Controller”

Facebook To Slurp Oculus Rift Users’ Every Move

The web is abuzz with the news that the Facebook-owned Oculus Rift has buried in its terms of service a clause allowing the social media giant access to the “physical movements and dimensions” of its users. This is likely to be used for the purposes of directing advertising to those users and most importantly for the advertisers, measuring the degree of interaction between user and advert. It’s a dream come true for the advertising business, instead of relying on eye-tracking or other engagement studies on limited subsets of users they can take these metrics from their entire user base and hone their offering on an even more targeted basis for peak interaction to maximize their revenue.

Hardly a surprise you might say, given that Facebook is no stranger to criticism on privacy matters. It does however represent a hitherto unseen level of intrusion into a user’s personal space, even to guess the nature of their activities from their movements, and this opens up fresh potential for nefarious uses of the data.

Fortunately for us there is a choice even if our community doesn’t circumvent the data-slurping powers of their headsets; a rash of other virtual reality products are in the offing at the moment from Samsung, HTC, and Sony among others, and of course there is Google’s budget offering. Sadly though it is likely that privacy concerns will not touch the non-tech-savvy end-user, so competition alone will not stop the relentless desire from big business to get this close to you. Instead vigilance is the key, to spot such attempts when they make their way into the small print, and to shine a light on them even when the organisations in question would prefer that they remained incognito.

Oculus Rift development kit 2 image: By Ats Kurvet – Own work, CC BY-SA 4.0, via Wikimedia Commons.

Insanely-Quick 3D Tracking with 1 Camera

Let’s face it: 3-dimensional odometry can be a computationally expensive problem often requiring expensive 3D cameras and optimized algorithms that can be difficult to wrap our head around. Nevertheless, researchers continue to push the bounds of visual odometry forward each year. This past year was no exception, as [Christian], [Matia], and [Davide] have tipped the scale in terms of speed with an algorithm that can track itself in 3D in real time.

In the video (after the break), the landmarks are sparse, the motion to track is relentlessly jagged, but SVO, or Semi-Fast Visual Odometry [PDF warning], keeps tracking its precision with remarkable consistency, making use of “high frequency texture” as a reference. Several other implementations require two cameras or a depth camera variant, but not SVO. It uses a single camera with a high frame rate between 55 and 300 frames per second. Best of all, the trio at the University of Zürich have made their codebase open source and available as a package for ROS.

Continue reading “Insanely-Quick 3D Tracking with 1 Camera”

DIY Virtual Reality Snowboard

If you’re looking for a quick and easy project to get into virtual reality, making your own VR skateboard controller is actually pretty easy to do!

First you’ll need some kind of VR headset. You could buy a fancy one, like the Oculus, or a Samsung Gear VR — or you could use something as simple as Google Cardboard — and you could even make your own. All it takes is a phone, an Arduino, a Bluetooth module, and an accelerometer-plus-gyroscope IMU.

Continue reading “DIY Virtual Reality Snowboard”

Wall-E Goes Corporate, Offering Telepresence Service

I guess if you are going to build a robot to do something boring like telepresence, you might as well make it cute. That’s obviously what [Andrew Maurer] was thinking when he built a telepresence robot using a Wall-E toy. The result is kind of adorable: Wall-E is holding the 5-inch HDMI screen that shows the video, and can scoot around in true Pixar fashion under remote control.

It’s also a neat build on the inside, using a Raspberry Pi for the brains and an Adafruit MotorHat to control the motors. The original toy didn’t have motors, so he added a new RC gearbox and motors to drive the little fella around. Installed behind Wall-Es eye is a USB webcam. Running behind the scenes is a mumble server that does the audio, a copy of Chromium that shows the video, and an Apache server that feeds the captured video to the other end of the conversation. The whole thing is tied together by a few scripts that kick things off appropriately and allow the user to remotely control Wall-E. It’s a cute build, and hopefully Wall-E can still find his EVE while performing his new corporate duties.

[via reddit]

Augmented Reality Ultrasound

Think of Virtual Reality and it’s mostly fun and games that come to mind. But there’s a lot of useful, real world applications that will soon open up exciting possibilities in areas such as medicine, for example. [Victor] from the Shackspace hacker space in Stuttgart built an Augmented Reality Ultrasound scanning application to demonstrate such possibilities.

But first off, we cannot get over how it’s possible to go dumpster diving and return with a functional ultrasound machine! That’s what member [Alf] turned up with one day. After some initial excitement at its novelty, it was relegated to a corner gathering dust. When [Victor] spotted it, he asked to borrow it for a project. Shackspace were happy to donate it to him and free up some space. Some time later, [Victor] showed off what he did with the ultrasound machine.

As soon as the ultrasound scanner registers with the VR app, possibly using the image taped to the scan sensor, the scanner data is projected virtually under the echo sensor. There isn’t much detail of how he did it, but it was done using Vuforia SDK which helps build applications for mobile devices and digital eye wear in conjunction with the Unity 5 cross-platform game engine. Check out the video to see it in action.

Thanks to [hadez] for sending in this link.

Continue reading “Augmented Reality Ultrasound”