When Hackaday announced winners of the 2014 Hackaday Prize, a bunch of hackers from Greece picked up the grand prize of $196,418 for their SatNOGS project – a global network of satellite ground stations for amateur Cubesats.
The design demonstrated an affordable ground station which can be built at low-cost and linked into a public network to leverage the benefits of satellites, even amateur ones. The social implications of this project were far-reaching. Beyond the SatNOGS network itself, this initiative was a template for building other connected device networks that make shared (and open) data a benefit for all. To further the cause, the SatNOGS team set up the Libre Space Foundation, a not-for-profit foundation with a mission to promote, advance and develop Libre (free and open source) technologies and knowledge for space.
Now, the foundation, in collaboration with the University of Patras, is ready to launch UPSat – a 2U, Open Source Greek Cubesat format satellite as part of the QB50 international thermosphere research mission. The design aims to be maximally DIY, designing most subsystems from scratch. While expensive for the first prototype, they hope that documenting the open source hardware and software will help kickstart an ecosystem for space engineering and technologies. As of now, the satellite is fully built and undergoing testing and integration. In the middle of July, it will be delivered to Nanoracks to be carried on a SpaceX Dragon capsule and then launched from the International Space Station.
Continue reading “After The Prize: SatNOGS Builds Satellites”
I still remember the first time I saw a satellite, I was 12 years old and was camping far away from the city lights. As I gazed up at the night sky, I could actually track satellites with my naked eye as they zoomed across the night’s sky. It was amazing. Nowadays, it’s getting harder to spot relatively small satellites with light pollution from large cities.
The International Space Station (ISS) on the other hand is a large piece of hardware — it’s about the size of a football field, and according to NASA it’s the second brightest object in the night sky. So why don’t we see it more often? Well, part of the reason is that you don’t know where to look. [Grady Hillhouse] set out to change that by building a what is basically a 2 degrees of freedom robot arm that will point you to where the ISS is at any given moment.
[Grady] uses a stepper motor for the azimuth, and a standard servo for the elevation, all powered by an Nucleo F401 development board, and an Adafruit motor shield and slip ring. The structure is made using some Erector set like parts from Actobotics.
He wrote the code from this open source project here. He’s currently cleaning up his code, and says he’ll be posting it up shortly. In the mean time, you can watch a video detailing the build in the video after the break. Or if you can’t wait, you can visit NASA’s web site to receive email or SMS messages on when the ISS is view-able in your hood.
Continue reading “It’s 10 PM, Do You Know Where Your Space Station Is At?”
According to ARISS (Amateur Radio on the International Space Station), the ISS will be sending us images using slow-scan TV on April 11th in honor of Russian cosmonaut Yuri Gagarin’s birthday. Tune in and you’ll get to see 12 different commemorative images from space, and of course bragging rights that you directly received them with your radio setup.
For those who aren’t Ham radio types, slow-scan TV (SSTV) is a radio mode where the pixels in an image are sent by encoding the brightness and/or color as a tone, a lot like a modem, fax machine, or the data cassette tapes of yore.
The ISS uses PD-180 which is a color mode where each pixel’s red, green, and blue values are encoded in a pitch between 1500 and 2300 Hz. Each image takes just over three minutes to transmit, meaning you’ll have to track the ISS pretty well as it travels across the sky. But don’t fret, they send each message for around an hour, so you have a good chance to receive it. (We’ll be the first to admit that a frame rate of one frame in 187 seconds isn’t really “TV”, but that’s what they call it.)
SSTV’s use in the space program goes back even before the moon landing, but with modern software-defined radio setups, it all becomes a lot more convenient to receive. The ISS folks do this periodically as a service to the amateur radio community, so it’s a good time to try out your chops.
We’ve covered ARISS before, but Yuri’s birthday is always a good reason to celebrate the folks out there. And if you need a reminder of when to look up, this hack right here has you covered.
If you do receive some images, you can upload them to the ARISS Gallery. Or you can just hit refresh to see them as others post them up.
Unless you’ve been living under a high voltage transformer, you’ve probably heard that NASA has grounded the Space Shuttle fleet. This makes getting stuff to and from the International Space Station slightly more difficult. With the growing need to get small experiments back to the surface quickly and safely, NASA is researching an idea they call Small Payload Quick Return, or SPQR (pdf warning). Basically, they toss the experiment out of the window, use drag to slow it down, and then use a High Altitude High Opening (HAHO) self guiding parafoil to steer the thing down to a predefined location on the surface.
Now, what we’re interested in is the self guided parafoil part, as it takes place in known hacker territory – around 100,000 feet. This is the altitude where most high altitude balloon experiments take place. NASA is throwing a bunch of money and brainpower to research this part of the system, but they’re having problems. Lots of problems.
Stick around after the break and see if you can help, and maybe pick up some ideas on how to steer your next High Altitude Balloon project back to the launch pad.
Continue reading “Ask Hackaday: Help NASA With Their High Altitude Problem”
The video above shows an animation of what the Canadian Space Agency hopes will be the first successful self-repair of the Mobile Servicing System aboard the ISS. The mobile servicing system is basically a group of several complicated robots that can either perform complicated tasks on their own, or be combined into a larger unit to extend the dexterity of the system as a whole.
The most recent addition to the servicing system is the Special Purpose Dexterous Manipulator, otherwise known as Dextre. Dextre is somewhat reminiscent of a human torso with two enormous arms. It is just one of the Canadian Space Agency’s contributions to the station. It was installed on the station in 2008 to perform activities that would normally require space walks. Dextre’s very first official assignment was successfully completed in 2011 when the robot was used to unpack two pieces for the Kounotori 2 transfer vehicle while the human crew on board the ISS was sleeping.
Dextre is constructed in such a way that it can be grabbed by the Canadarm2 robot and moved to various work sites around the Space Station. Dextre can then operate from the maintenance site on its own while the Canadarm2 can be used for other functions. Dextre can also be operated while mounted to the end of Canadarm2, essentially combining the two robots into one bigger and more dexterous robot.
One of the more critical camera’s on the Canadarm2 has started transmitting hazy images. To fix it, the Canadarm2 will grab onto Dextre, forming a sort of “super robot”. Dextre will then be positioned in such a way that it can remove the faulty camera. The hazy camera will then be mounted to the mobile base component of the Mobile Servicing System. This will give the ISS crew a new vantage point of a less critical location. The station’s human crew will then place a new camera module in Japan’s Kibo module’s transfer airlock. Dextre will be able to reach this new camera and then mount it on the Canadarm2 to replace the original faulty unit. If successful, this mission will prove that the Mobile Servicing System has the capability to repair itself under certain conditions, opening the door for further self-repair missions in the future.
The first element of the International Space Station (ISS) launched over fifteen years ago, on November 20, 1998. For more than thirteen years at least two human beings have been continually living off the surface of our planet. Assembly of the Space Station is now complete. It is being utilized by its crews and scientists from around the world to execute its primary mission – scientific investigations that can only be accomplished in the microgravity environment of Low Earth Orbit (LEO). As with any structure, items age, wear out, or break and need to be repaired. What could be rather “simple” repairs on Earth can become much more complex in zero gravity. In some cases, “necessity becomes the mother of invention.”
Continue reading “The Pioneering Lifestyle in Low Earth Orbit”
If you live in the Eastern portion of the United States and the skies are clear you can see a student built satellite flashing LEDs in Morse Code today. But don’t worry. If you it’s cloudy or if you live elsewhere there are several other opportunities to see it in the coming days.
This is the Niwaka Fitsat-1. It was developed by students at the [Fukuoka Institute of Technology] and deployed from the International Space Station on October 4th. Included in the payload is an array of LEDs seen in the image above. On a set schedule these are used to flash a Morse Code message for two minutes at a time. That is what’s shown in the image on the upper right.
You can look up information on seeing Fitsat-1 in your own area using this webpage. All of the observation windows in our area require a pair of binoculars or better. We’re not sure if there is any case in which this can be seen by the naked eye.
[Thanks SWHarden and KomradBob]