There are very few things that are surrounded with as much hearsay and rumor as the origins of the QWERTY layout of typewriters and keyboards. The reason behind the QWERTY layout isn’t as simple as ‘so the bars for each letter don’t collide with each other.’ That’s nonsense – it would make far more sense to improve the mechanism before changing the arrangement of the keyboard around.
That’s not the only fallacious argument for the creation of QWERTY. It’s also been called a marketing ploy; Stephen Jay Gould popularized the idea of the QWERTY keyboard being as it is so a salesman could peck out TYPE WRITER on the top row . This also makes little sense. Why would the top row and not the home row be so privileged as to contain all the letters the make up the name of the machine. For that matter, wouldn’t a sales pitch be more impressive if TYPE WRITER were typed with one hand?
This doesn’t mean there’s not a method behind the madness of QWERTY – it’s just not as simple as jammed typewriter mechanisms or appeasing the wishes of salesmen in the 1870s. QWERTY didn’t come out of thin air, though, but folk tale history of this keyboard layout is sadly deficient.
If you own a video projector, be it a module small enough to fit in a mobile phone or one designed for a cinema screen, the chances are it will have a DLP at its heart. An array of microscopic mirrors on an integrated circuit, the current state of the art in video projection technology.
Perhaps you own an older video projector, or maybe a cheaper new one. If so the chances are it’ll have a small LCD screen doing its work, taking the place of the Kodachrome in something very similar to your grandparents’ slide projector or their grandparents’ magic lantern.
LCD technology was invented in the 1970s, while DLP was invented at the end of the 1980s. So how did the video projectors that were such a staple of televised spectaculars in the preceding decades work? For that matter, how did NASA project their status displays on the huge screen at Mission Control? Certainly not with CRT technology, even the brightest CRT projectors weren’t up to filling a cinema-sized screen.
The answer came from the Eidophor (Greek: ‘eido’ and ‘phor’, ‘image’ and ‘bearer’), a device invented in the years before World War II by the Swiss physicist Dr. Fritz Fischer and granted a US patent in 1945. It featured a complex vacuum device in which an electron gun painted the video frames as a raster on an oil-covered mirror in the light path of a fairly conventional projector. High-voltage electric charges have the effect of deforming the surface of mineral oils, and it was this effect that was exploited to vary the effectiveness of the mirror as the raster was drawn. An unfortunate side-effect of tracing an oil surface with an electron beam is that a charge will build up on the oil surface, so the entire oil-covered mirror assembly had to rotate within its vacuum enclosure and pass under an electrode which removed any charge build-up.
You will probably be unaware of the exact date you last saw an eidophor performance. Quince Imaging tell us their last one was used at the TWA Dome in St Louis in July 2000. Eidophores may have become more compact over the decades but they remained costly to run, and through the 1990s they were suplanted by DLP devices that did substantially the same job with a lot less fuss.
It is not often that a search in the Hackaday archives for a technology returns no results, but the eidophor is one of those cases. Perhaps that is a fitting epitaph for a device that created its own show but never starred in it, that it is only its spectacular performances that live on.
We often see “logic analyzer” projects which are little more than microcontrollers reading data as fast as they can, sending it to a PC, and then plotting the results. Depending on how fast the microcontroller is, these projects range from adequate to not very useful.
At first glance, [esot.eric’s] logic analyzer project has an AVR in it, so it ought to be on the low end of the scale. Then you look at the specs: 32 channels at 30 megasamples per second. How does that work with an AVR in it?
The answer lies in the selection of components. The analyzer uses a 128MB SDRAM DIMM (like an older PC might use for main memory). That makes sense; the Arduino can’t store much data internally. However, it isn’t the storage capacity that makes this choice critical. It seems [esot.eric] has a way to make the RAM “free run”.
The idea is to use the Arduino (or other host microcontroller) to set up the memory. Some of the memory’s output bits feedback to the address and data lines. Then the microcontroller steps aside and the SDRAM clocks samples into its memory by itself at the prevailing clock rate for the memory.