When you get into making videos of products or your own cool hacks, at some point you’re going to start wondering how those neat panning and rotating shots are achieved. The answer is quite often some kind of mechanical slider which sends the camera along a predefined path. Buying one can be an expensive outlay, so many people opt to build one. [Rahel zahir Ali] was no different, and designed and built a very simple slide, but with a neat twist.
This design uses a geared DC motor, taken from a car windscreen wiper. That’s a cost effective way to get your hands on a nice high-torque motor with an integral reduction gearbox. The added twist is that the camera mount is pivoted and slides on a third, central smooth rod. The ends of this guide rod can be offset at either end, allowing the camera to rotate up to thirty degrees as the slide progresses from one end to the other. With a few tweaks, the slider can be vertically mounted, to give those up-and-over shots. Super simple, low tech and not an Arduino in sight.
The CAD modelling was done with Fusion 360, with all the models downloadable with source, in case someone needs to adapt the design further. We were just expecting a pile of STLs, so seeing the full source was a nice surprise, given how many open source projects like this (especially on Thingiverse) do often seem to neglect this.
Electronics consist of a simple DC motor controller (although [Rahel] doesn’t mention a specific product, it should not be hard to source) which deals with the speed control, and a DPDT latching rocker switch handles the motor direction. A pair of microswitches are used to stop the motor at the end of its travel. Other than a 3D printer, there is nothing at all special needed to make yourself quite a useful little slider!
[T-Kuhn]’s Octo-Bouncer platform has learned some new tricks since we saw it last. If you haven’t seen it before, this device uses computer vision from a camera mounted underneath its thick, clear acrylic platform to track a ball in 3D space, and make the necessary (and minute) adjustments needed to control the ball’s movements with a robotic platform in real time.
We loved the Octo-Bouncer’s mesmerizing action when we saw it last, and it’s only gotten better. Not only is there a whole new custom ball detection algorithm that [T-Kuhn] explains in detail, there are also now visualizations of both the ball’s position as well as the plate movements. There’s still one small mystery, however. Every now and again, [T-Kuhn] says that the ball will bounce in an unexpected direction. It doesn’t seem to be a bug related to the platform itself, but [T-Kuhn] has a suspicion. Since contact between the ball and platform is where all the control comes from, and the ball and platform touch only very little during a bounce, it’s possible that bits of dust (or perhaps even tiny imperfections on the ball’s surface itself) might be to blame. Regardless, it doesn’t detract from the device’s mesmerizing performance.
Design files and source code are available on the project’s GitHub repository for those who’d like a closer look. It’s pretty trippy watching the demonstration video because there is so much going on at once; you can check it out just below the page break.
If you’ve ever watched one of those high production value YouTube videos and wondered how they’re able to get those smooth shots where the camera seems to be spinning around an object, you were probably looking at the product of an motorized camera motion system. There’s no question these rigs can produce visually striking shots, but their high cost usually keeps them out of the hands of us lowly hackers.
Unless of course you do like [Andy], and build your own. The latest version of this impressive rig features the ability to continuously rotate thanks to commercial 12-wire slip rings, with optical endstops so the machine can still be homed at the beginning of a move. An onboard Raspberry Pi and Arduino Uno are responsible for controlling the stepper motors, the configuration of which ends up being reminiscent of a standard 3D printer.
The software [Andy] has come up with lets him synchronize the camera rig with a small rotating platform he built, which allows for even more complex shots as demonstrated in the video below. It also supports a very slick MQTT-enabled remote controller that he built as a previous project, which makes taking direct control over the camera and monitoring its status much easier.
DJI recently introduced a slick motion controller that eschews the traditional dual-stick transmitter and allows you to fly their new “FPV Drone” with just one hand. The fact that it looks like it could double as the control stick for an X-Wing is just an added bonus. Unfortunately, that single model is the only thing the $199 USD controller is currently compatible with. Unwilling to get locked into the DJI ecosystem, [Paweł Spychalski] has developed an open source work-alike motion controller that brings gesture flying to home-built quadcopters and airplanes.
Now to be clear, you’ll still need a traditional transmitter to use this device. Rather than trying to reinvent the wheel, [Paweł] decided to implement his motion controller as an add-on for OpenTX hardware like the RadioMaster TX16S. It simply plugs into the trainer port on the back of the transmitter and acts as a secondary input. This greatly simplifies the design, as it essentially just needs to read angle data from its MPU-6050 gyro/accelerometer and forward it along to OpenTX over the serial port. Plus the fact that it’s connected to the trainer port means you can disable it and return to traditional controls in an instant if anything goes wrong.
Outside of the motion sensing gear, the ESP32-powered peripheral also has a thumb stick and a pair of push buttons nestled into its 3D printed frame. An OLED display provides some user feedback, and a holder for a 18650 cell is mounted to the back side as the controller will need its own power source when [Paweł] gets around to making its connection to the transmitter wireless.
In the video below, [Paweł] takes the motion controller for a test flight and comes away largely satisfied with the results. Some tweaks are in the works as you might expect for a first attempt, but nothing that would prevent you from building your own version today and experiencing what might be the next evolution of RC flying.
“Scope creep” is often derided as an obstacle between your idea and the delivery of a finished project. That may be, but sometimes the creep is the whole point. It’s how we end up with wonderful builds like this multi-axis differential camera slider.
We mention scope creep because that’s what [Jan Derogee] blames for this slider’s protracted development time, as well as its final form. The design is a bit unconventional in that it not only dollies the camera left and right but also works in pan and tilt axes, and it does this without putting any motors on the carriage. Instead, the motors, which are located near the end of the slider rails, transmit power to the carriage via loops of 217timing belt. It’s a little like the CoreXY mechanism; rotating the motors in the same direction and speed slides the carriage, while moving them in opposite directions pans the camera. A Sparkfun Pro Micro in the controller coordinates the motors for smooth multi-axis motion, and the three steppers — there’s a separate motor for the tilt axis — sound really cool all working at the same time. Check out the video below for the full story.
The rig consists of a series of 3D printed axes all joined together into a 6-axis motion rig. Additionally, actuators attached to the lens of the camera allow zoom and focus to be be controlled programmatically too. An Arduino runs the show, interpreting G-code and running the various axes, with a Raspberry Pi acting as a gateway to allow the rig to be commanded from PCs or smartphones.
Currently, control is largely manual, by entering G-code commands to move the rig in various ways. The rig can also have its motors temporarily disengaged by a button, allowing the camera to be aimed by hand, before holding the position. In this way, it acts as a highly versatile tripod. Future plans involve more automation if suitable open-source software can be found.
While there’s been a lot of advancements in VR gaming over the last couple of years, plenty of folks are still happy enough to just stare at their monitor. But that’s not to say some of those fancy head-tracking tricks wouldn’t be a welcome addition to their repertoire. For players who are literally looking to get their head in the game, [Adrian Schwizgebel] has created qeMotion.
The idea here is simple enough: attach a motion sensor to a standard gaming headset (here a MPU-6050 IMU), and use the data from it to virtually “press” keys through USB HID emulation. Many first person shooter games offer the ability to lean left or right by pressing Q or E respectively, so all [Adrian] had to do was map the appropriate accelerometer readings to those keys for it to work seamlessly with popular titles such as Tom Clancy’s Rainbow Six Siege and Insurgency.
The concept might be basic, but the execution is anything but. Rather than just duct taping an Arduino to his headset, [Adrian] designed a very slick 3D printed enclosure for the electronics that sits on his desk. While they haven’t all been implemented yet, the devices features indicator lights and buttons to switch through various modes. The sensor on the headset has similarly been encased in a very professional looking 3D printed box, complete with a nice braided cable to link it to the desk unit.