Spacecraft rocket engines come in a variety of forms and use a variety of fuels, but most rely on chemical reactions to blast propellants out of a nozzle, with the reaction force driving the spacecraft in the opposite direction. These rockets offer high thrust, but they are relatively fuel inefficient and thus, if you want a large change in velocity, you need to carry a lot of heavy fuel. Getting that fuel into orbit is costly, too!
Ion thrusters, in their various forms, offer an alternative solution – miniscule thrust, but high fuel efficiency. This tiny push won’t get you off the ground on Earth. However, when applied over a great deal of time in the vacuum of space, it can lead to a huge change in velocity, or delta V.
This manner of operation means that an ion thruster and a small mass of fuel can theoretically create a much larger delta-V than chemical rockets, perfect for long-range space missions to Mars and other applications, too. Let’s take a look at how ion thrusters work, and some of their interesting applications in the world of spacecraft!
We have to confess that we occasionally send friends a link to “let me Google that for you” when they ask us something that they could have easily found online. Naturally, if someone asked us how big the moon is, we’d ask Google or another search engine. But not [Prof Matt Strassler]. He’d tell you to figure it out yourself and he would then show you how to do it.
This isn’t a new question. People have been wondering about the moon since the dawn of human civilization. The ancient Greeks not only asked the question, but they worked out a pretty good answer. They knew approximately how big the Earth was and they knew the moon was far away because it is seen over a very wide area. They also knew the sun was even further away because the moon sometimes blocks the sun’s light in an eclipse. Using complex geometry and proto-trigonometry they were able to work out an approximate size of the moon. [Matt’s] method is similar but easier and relies on the moon occluding distant stars and planets.
Though some of us are heavily assisted by smart phone apps and delivery, humans don’t need GPS to find food. We know where the fridge is. The grocery store. The drive-thru. And we don’t really need a map to find shelter, in the sense that shelter is easily identifiable in a storm. You might say that our most important navigation skills are innate, at least when we’re within our normal environment. Drop us in another city and we can probably still identify viable overhangs, cafes, and food stalls.
The question is, do these navigational skills vary by species or environment? Or are the tools necessary to forage for food, meet mates, and seek shelter more universal? To test the waters of this question, Israeli researchers built a robot car and taught six fish to navigate successfully toward a target with a food reward. This experiment is one of domain transfer methodology, which is the exploration of whether a species can perform tasks outside its natural environment. Think of all the preparation that went into Vostok and Project Mercury.
The principles of open-source hardware are starting to make great strides in scientific research fields. [Walker Arce] tells us about his paper co-authored with [Jeffrey R. Stevens], about a dog treat dispenser designed with scientific researchers in mind – indispensable for behavior research purposes, and easily reproducible so that our science can be, too. Use of Raspberry Pi, NEMA steppers and a whole lot of 3D printed parts make this build cheap (< $200 USD) and easy to repeat for any experiments involving dogs or other treat-loving animals.
Even if you’re not a scientist, you could always build one for your own pet training purposes – this design is that simple and easy to reproduce! The majority of the parts are hobbyist-grade, and chances are, you can find most of the parts for this around your workshop. Wondering how this dispenser works, and most importantly, if the dogs are satisfied with it? Check out a short demonstration video after the break.
Despite such dispensers being commercially available, having a new kind of dispenser designed and verified is more valuable than you’d expect – authors report that, in their experience, off-the-shelf dispensers have 20-30% error rate while theirs can boast just 4%, and they have test results to back that up. We can’t help but be happy that the better-performing one is available for any of us to build. The GitHub repository has everything you could want – from STLs and PCB files, to a Raspberry Pi SD card image and a 14-page assembly and setup guide PDF.
The eyes are windows into the mind, and this research into what jumping spiders look at and why required a clever device that performs eye tracking, but for jumping spiders. The eyesight of these fascinating creatures in some ways has a lot in common with humans. We both perceive a wide-angle region of lower visual fidelity, but are capable of directing our attention to areas of interest within that to see greater detail. Researchers have been able to perform eye-tracking on jumping spiders, literally showing exactly where they are looking in real-time, with the help of a custom device that works a little bit like a miniature movie theatre.
To do this, researchers had to get clever. The unblinking lenses of a spider’s two front-facing primary eyes do not move. Instead, to look at different things, the cone-shaped inside of the eye is shifted around by muscles. This effectively pulls the retina around to point towards different areas of interest. Spiders, whose primary eyes have boomerang-shaped retinas, have an X-shaped region of higher-resolution vision that the spider directs as needed.
So how does the spider eye tracker work? The spider perches on a tiny foam ball and is attached — the help of a harmless and temporary adhesive based on beeswax — to a small bristle. In this way, the spider is held stably in front of a video screen without otherwise being restrained. The spider is shown home movies while an IR camera picks up the reflection of IR off the retinas inside the spider’s two primary eyes. By superimposing the IR reflection onto the displayed video, it becomes possible to literally see exactly where the spider is looking at any given moment. This is similar in some ways to how eye tracking is done for humans, which also uses IR, but watches the position of the pupil.
In the short video embedded below, if you look closely you can see the two retinas make an X-shape of a faintly lighter color than the rest of the background. Watch the spider find and focus on the silhouette of a tasty cricket, but when a dark oval appears and grows larger (as it would look if it were getting closer) the spider’s gaze quickly snaps over to the potential threat.
It is my pleasure to announce that Keith Thorne has graciously agreed to deliver a keynote take at Hackaday Remoticon 2. Get your ticket now!
Keith is an astrophysicist and has worked on the Laser Interferometer Gravitational-Wave Observatory (LIGO) since 2008, literally looking for ripples in space-time that you know as gravitational waves. The effects of the phenomena are so subtle that detecting an event requires planet-scale sensors in the form of 4 km long interferometers placed in different parts of the United States whose readings can be compared against one another. A laser beam inside these interferometers bounces back and forth 300 times for a total travel distance of 1,200 km in which any interaction with gravitational waves will ever-so-slightly alter how the photons from the beam register.
The challenges of building, operating, and interpreting such a device are manifold. These interferometers are the highest precision devices ever devised, able to detect motion 1/10,000 of the diameter of a proton! To get there, the mirrors need to be cooled to 77 nano-Kelvins. Getting the most out of it is what Keith and the rest of the team specialize in. This has included things like hacking the Linux kernel to achieve a sufficient level of real-time digital control, and using “squeezed light” to improve detection sensitivity in frequencies where quantum mechanics is getting in the way. While the detectors were first run in 2015 & 2016, successfully observing three events, the work to better understand this phenomenon is ongoing and may include a third site in India, and a space-based detector in the future.
In getting to know Keith he mentioned that he is excited to speak to a conference packed with people who want to hear the gory technical details of this fantastic piece of hardware. I’m sure we’re all giddy to learn what he has to say. But if you’re someone who wants to work on a project like this, he tipped us off that there’s an active EE job posting for LIGO right now. Maybe you’ll end up doing the keynote at a future Hackaday conference.
Call for Proposals is Still Open!
We’re still on the hunt for great talks about hardware creation. True creativity is fed by a steady stream of inspiration. Be that inspiration by giving a talk about the kinds of things you’ve been working on!
Liquid-fuelled rocket engine design has largely followed a simple template since the development of the German V-2 rocket in the middle of World War 2. Propellant and oxidizer are mixed in a combustion chamber, creating a mixture of hot gases at high pressure that very much wish to leave out the back of the rocket, generating thrust.
Humans love combusting fuels in order to do useful work. Thus far in our history, whether we look at steam engines, gasoline engines, or even rocket engines, all these technologies have had one thing in common: they all rely on fuel that burns in a deflagration. It’s the easily controlled manner of slow combustion that we’re all familiar with since we started sitting around campfires. Continue reading “Japanese Rocket Engine Explodes: Continuously And On Purpose”→