If your only exposure to seismologists at work is through film and television, you can be forgiven for thinking they still lay out rolls of paper to examine lines of ink under a magnifying glass. The reality is far more interesting in a field that has eagerly adopted all available technology. A dramatic demonstration of modern earthquake data gathering, processing, and visualization was Tweeted by @IRIS_EPO following a central California quake on July 4th, 2019. In this video can see the quake’s energy propagate across the continental United States in multiple waves of varying speed and intensity. The video is embedded below, but click through to the Twitter thread too as it has a lot more explanation.
The acronym IRIS EPO expands out to Incorporated Research Institutions for Seismology, Education and Public Outreach. We agree with their publicity mission; more people need to know how cool modern seismology is. By combining information from thousands of seismometers, we could see forces that we could not see from any individual location. IRIS makes seismic data available to researchers (or curious data science hackers) in a vast historical database or a real time data stream. Data compilations are presented in several different forms, this particular video is a GMV or Ground Motion Visualization. Significant events like the 4th of July earthquake get their own GMV page where we can see additional details, like the fact this visualization compiled data from 2,132 stations.
The finished product shows very well the progression as the youngster adapts to a regular sleep pattern, and even shows a shift to the right at the very bottom as a result of a trip across time zones to see relatives. It’s both a good visualisation and a unique keepsake that the baby will treasure one day as an adult. (Snarky Ed Note: Or bring along to the therapist as evidence.)
March 14th is “Pi Day”, for reasons which should be obvious to our more mathematically inclined readers. As you are not reading this post on March 14th, that must mean we’re either fashionably late to Pi Day 2019, or exceptionally early for Pi Day 2020. But in either event, we’ve got a hack for you that celebrates the day using two things we have it on good authority most hackers overindulge in: food and needless complexity.
One might say that this project takes a three dimensional pie and converts it to a two dimensional facsimile, but perhaps that’s over-analyzing it. In reality, it was a fun little hack [Mike] put together just because he thought it would be fun. Which is certainly enough of a motive for us. More practically though, if you’re looking for a good example for how to get a load cell talking to your non-edible Raspberry Pi, you could do worse than checking this out.
Like many people who have a solar power setup at home, [Jeroen Boeye] was curious to see just how much energy his panels were putting out. But unlike most people, it just so happens that he’s a data scientist with a deep passion for programming and a flair for visualizations. In his latest blog post, [Jeroen] details how his efforts to explain some anomalous data ended with the discovery that his solar array was effectively acting as an extremely low-resolution camera.
It all started when he noticed that in some months, the energy produced by his panels was not following the expected curve. Generally speaking, the energy output of stationary solar panels should follow a clear bell curve: increasing output until the sun is in the ideal position, and then decreasing output as the sun moves away. Naturally cloud cover can impact this, but cloud cover should come and go, not show up repeatedly in the data.
[Jeroen] eventually came to realize that the dips in power generation were due to two large trees in his yard. This gave him the idea of seeing if he could turn his solar panels into a rudimentary camera. In theory, if he compared the actual versus expected output of his panels at any given time, the results could be used as “pixels” in an image.
He started by creating a model of the ideal energy output of his panels throughout the year, taking into account not only obvious variables such as the changing elevation of the sun, but also energy losses through atmospheric dispersion. This model was then compared with the actual power output of his solar panels, and periods of low efficiency were plotted as darker dots to represent an obstruction. Finally, the plotted data was placed over a panoramic image taken from the perspective of the solar panels. Sure enough, the periods of low panel efficiency lined up with the trees and buildings that are in view of the panels.
If you are an astronomy buff, there are plenty of star maps you can find in print or online (or even on your Smartphone). But if you are a science fiction fan (or writer), you probably find those maps frustrating because they are flat. Two stars next to each other on the map might be light years apart in the axis coming out of the page. A star 3.2 light years from Sol (our sun) looks the same on the map as a star 100 light years away.
The Gaia satellite (an ESA project) orbits beyond the moon and is carefully mapping the 3D position of every point of light it sees. [Charlie Hoey] took the data for about 2 million stars and used WebGL to give you a 3D view of the data in your web browser.