Smart Thermostats Pitched For Texas Homes To Relieve Stressed Grid

It’s not much of a secret that Texas’ nearly completely isolated grid is in a bit of a pickle, with generating capacity often being handily outstripped during periods of extreme demand. In a latest bid to fight this problem, smart thermostats are being offered to customers, who will then participate in peak-shaving. The partnership between NRG Energy Inc., Renew Home LLC, and Alphabet Inc. will see about 650,000 of these thermostats distributed to customers.

For customers the incentive would be mostly financial, though the details on the potential cost savings seem scarce. The thermostats would be either a Vivint (an NRG company) or Google Nest branded one, which would be controlled via Google Cloud, allowing for thermostat settings to be changed to reduce the load on the grid. This is expected to save ‘300 MW’ in the first two years, though it’s not clear whether this means ‘continuously’, or intermittent like with a peaker natural gas plant.

Demand curtailment is not a new thing, with it being a big thing among commercial customers in South Korea, as we discussed within the topic of vehicle-to-grid energy storage. Depending on how it is implemented it can make a big difference, but it’ll remain to see how regular consumers take to the idea. It also provides more evidence for reducing grid load being a lot easier than adding grid-level storage, which is becoming an increasingly dire topic as more non-dispatchable solar and wind power is added to the grid.

Building A Reproduction Apple I

If you think of Apple today, you probably think of an iPhone or a Mac. But the original Apple I was a simple PC board and required a little effort to start up a working system. [Artem] has an Apple I reproduction PCB, and decided to build it on camera so we could watch.

For the Apple I, the user supplied a keyboard and some transformers, so [Artem] had to search for suitable components. He wisely checks the PCB to make sure there are no shorts in the traces. From there, you can watch him build the machine, but be warned: even with speed ups and editing, the video is over an hour long.

If you want to jump to the mostly working device, try around the 57-minute mark. The machine has a basic ROM monitor and, of course, needs a monitor. There was a small problem with memory, but he eventually worked it out by inhibiting some extra RAM on the board. Troubleshooting is half of the battle getting something like this.

Want to look inside the clock generator chip? Or skip the PCB and just use an FPGA.

Continue reading “Building A Reproduction Apple I”

Laser Sound Visualizations Are Not Hard To Make

You might think that visualizing music with lasers would be a complicated and difficult affair. In fact, it’s remarkably simple if you want it to be, and [byte_thrasher] shows us just how easy it can be.

At heart, what you’re trying to do is make a laser trace out waveforms of the music you’re listening to, right? So you just need a way to move the laser’s beam along with the sound waves from whatever you’re listening to. You might be thinking about putting a laser on the head of a servo-operated platform fed movement instructions from a digital music file, but you’d be way over-complicating things. You already have something that moves with the music you play — a speaker!

[byte_thrasher’s] concept is simple. Get a Bluetooth speaker, and stick it in a bowl. Cover the bowl with a flexible membrane, like plastic wrap. Stick a small piece of mirror on the plastic. When you play music with the speaker, the mirror will vibrate and move in turn. All you then have to do is aim a safe laser in a safe direction such that it bounces off the mirror and projects on to a surface. Then, the laser will dance with your tunes, and it’ll probably look pretty cool!

We’ve seen some beautiful laser visual effects before, too. Just be careful and keep your power levels safe and your beams pointing where they should be.

Continue reading “Laser Sound Visualizations Are Not Hard To Make”

AI Face Anonymizer Masks Human Identity In Images

We’re all pretty familiar with AI’s ability to create realistic-looking images of people that don’t exist, but here’s an unusual implementation of using that technology for a different purpose: masking people’s identity without altering the substance of the image itself. The result is the photo’s content and “purpose” (for lack of a better term) of the image remains unchanged, while at the same time becoming impossible to identify the actual person in it. This invites some interesting privacy-related applications.

Originals on left, anonymized versions on the right. The substance of the images has not changed.

The paper for Face Anonymization Made Simple has all the details, but the method boils down to using diffusion models to take an input image, automatically pick out identity-related features, and alter them in a way that looks more or less natural. For this purpose, identity-related features essentially means key parts of a human face. Other elements of the photo (background, expression, pose, clothing) are left unchanged. As a concept it’s been explored before, but researchers show that this versatile method is both simpler and better-performing than others.

Diffusion models are the essence of AI image generators like Stable Diffusion. The fact that they can be run locally on personal hardware has opened the doors to all kinds of interesting experimentation, like this haunted mirror and other interactive experiments. Forget tweaking dull sliders like “brightness” and “contrast” for an image. How about altering the level of “moss”, “fire”, or “cookie” instead?

A light grey box about the size of a brick with exposed screws held in a person's hand. There are two illuminated push buttons on the bottom left of the top panel. One is illuminated blue while the other is green. A small square screen sits next to a bank of nine different sections with an LED indicator and text of "HW, BAT, HBEAT, ECG, LOD +, LOD -, PPG, Pump, Valve."

Open Cardiography Signal Measuring Device

Much of the world’s medical equipment is made by a handful of monopolistic megacorps, but [Milos Rasic] built an open cardiography signal measuring device for his master’s thesis.

Using a Pi Pico W for the brains, [Rasic]’s device can record, store and analyze the data from an arm cuff, stethoscope, electrocardiograph (ECG), and pulse oximeter. This data can be used for monitoring blood pressure in patients and he has results from some of his experiments to determine the optimal algorithm for the task on the GitHub if you really want to get into the nitty gritty details.

Inside the brick-sized enclosure is the custom PCB, an 18650 Li-ion cell, and a pneumatic assembly for the arm cuff. Medical sensors attach via GX12 connectors on the back, a USB type B connector is used for data, and a USB C connector provides power for the device. The brightly colored labels will no doubt come in handy in a clinical setting where you really want to be sure you’ve got everything plugged in correctly.

Want more open medical equipment? How about an open ECG or this less accurate, but more portable, credit card ECG? We’d be remiss not to mention the huge amount of work on ventilators during the worst days of the COVID-19 pandemic as well.

Landscape Motif Makes This E-Ink Weather Display Easy To Understand

True weather geeks will disagree, but there might be a better way to know how to dress for the day than divining what the weather will likely be from the current readings for temperature, pressure, humidity, and wind. Sure, the data will give you a good idea of where the weather is heading, but perhaps a quick visual summary such as the one offered by this pictorial landscape weather display is a better way to get out the door in the morning.

While many consumer weather stations incorporate some kind of graphical forecast for quick reference, [lds133] took a slightly different approach to forecasting. A cartoon landscape represents the day ahead, with various elements representing the coming weather scrolling across the display as time progresses. Trees are used to indicate wind direction and speed, with palm trees indicating south wind and pine trees winds from the north, and the taller the trees, the stronger the wind. The forest floor rises and falls with the expected temperature, the sun and moon appear at the proper time to indicate sunrise and sunset, and cloud icons are added when needed to show the degree of cloud cover. And because into each life a little rain must fall, animations show when you can expect rain or snow.

As for the electronics, if you think this would be a perfect application for an E-ink module, [lds133] agrees. The 296×128 pixel Waveshare display is the perfect aspect ratio for the job and provides nice, crisp icons. The display is updated every 15 minutes from the OpenWeather API by a Python program running on an ESP32 behind the scenes.

We’ve seen similar graphical forecast displays before, but we get it if that’s not your thing. Perhaps a more data-driven weather forecast will suit you better?

FLOSS Weekly Episode 809: Pi4J – Stable And Boring On The Raspberry Pi

This week, Jonathan Bennett and David Ruggles chat with Frank Delporte about Pi4J, the friendly Java libraries for the Raspberry Pi, that expose GPIO, SPI, I2C and other IO interfaces. Why would anyone want to use Java for the Pi? And what’s changed since the project started? Listen to find out!

Continue reading “FLOSS Weekly Episode 809: Pi4J – Stable And Boring On The Raspberry Pi”