Apple Archeology: The Future Once Had Server Side Computing In It

To read the IT press in the early 1990s, those far-off days just before the Web was the go-to source of information, was to be fed a rosy vision of a future in which desktop and server computing would be a unified and powerful experience. IBM and Apple would unite behind a new OS called Taligent that would run Apple, OS/2, and 16-bit Windows code, and coupled with UNIX-based servers, this would revolutionise computing.

We know that this never quite happened as prophesied, but along the way, it did deliver a few forgotten but interesting technologies. [Old VCR] has a look at one of these, a feature of the IBM AIX, which shipped with mid-90s Apple servers as a result of this partnership, in which Mac client applications could have server-side components, allowing them to offload computing power to the more powerful machine.

The full article is very long but full of interesting nuggets of forgotten 1990s computing history, but it’s a reminder that DOS/Windows and Novell Netware weren’t the only games in town. The Taligent/AIX combo never happened, but its legacy found its way into the subsequent products of both companies. By the middle of the decade, even Microsoft had famously been caught out by the rapid rise of the Web. He finishes off by creating a simple sample application using the server-side computing feature, a native Mac OS application that calls a server component to grab the latest Hacker News stories. Unexpectedly, this wasn’t the only 1990s venture from Apple involving another company’s operating system. Sometimes, you just want to run Doom.

Impostor Syndrome: It’s Not Your Fault!

[Crispernaki] and I have something in common. We both saw this awesome project that made a scroll wheel out of a VHS head back in 2010, and wanted to make one. We both wanted to put our own spin on the gadget, (tee-hee), discovered that it was harder than either of us wanted to commit to, and gave up.

Flash forward about a million Internet years, and [crispernaki] finally made his and wrote it up. The only problem is that it was too easy. In 2010, making USB gadgets was a lot more involved than it is today. (Back then, we had to chisel device descriptors on stone tablets.) Nowadays, the firmware is just a matter of importing the right library, and the hardware is a magnetic rotation sensor breakout board, a magnet, and super glue. Cheap, and easy.

All of this led our hero to feeling insecure. After all, a hack that beat him a dozen years ago turned out to be dead easy today. Maybe it was too easy? Maybe he wasn’t a “real” hacker? These are the signs of impostor syndrome – that feeling that just because you aren’t the world’s best, or climbing the highest mountain, or hacking the hardest project, you’re not worthy.

Well, listen up. Impostors don’t finish projects, and impostors don’t write them up to share with all the rest of us. By actually doing the thing – hacking the hack – all chances of being a fake are ruled out. The proof is sitting there on your desk, in all its Altoids-tin glory.

And it’s not your fault that it was too easy this time around. You can’t do anything to turn back the hands of time, to make the project any harder these days, or to undo the decade of hacker technical progress on the software side, much less change the global economy to make a magnetic sensor unobtainable again. The world improved, you got your hack done, and that’s that. Congratulations! (Now where do I buy some of those on-axis magnets?)

Illustrated Kristina with an IBM Model M keyboard floating between her hands.

Keebin’ With Kristina: The One With The Foot Keyboard

[crispernaki]’s opening comments to this VCR head scroll wheel project lament that overall technical details aren’t “complex, ground-breaking, or even exciting.” Since when does that matter? The point is that not only did the thing finally, eventually get built, it gets daily use and it sparks joy in its owner.

This feel-good story is one of procrastination, laziness, and one aha! moment, and it’s roughly twelve years in the making. Inspired by an Instructable from long ago, [crispernaki] ran straight to the thrift store to get a VCR and take it apart.

The original plan was to just reuse the VCR head’s PCB and hide it in an enclosure, and then figure out way to block and unblock the path between an IR emitter/receiver pair. After many disemboweled mice and fruitless attempt, the project was once again shelved.

But then, [crispernaki] remembered the magnetic rotary encoder demo board that was just sitting around, along with various microcontrollers and Altoids tins. And it all quickly came together with a Teensy 2.0 and some bits and bobs, including a magnet glued on the shaft of the VCR head. A chip on the demo board does all the heavy lifting, and of course, the Teensy does the work of emulating an HID.

Continue reading “Keebin’ With Kristina: The One With The Foot Keyboard”

Atari Gets Semi-Modern Video Output

The Atari 2600 is a historical enigma in many ways. On one hand, it was the most popular gaming console of its era, but it was also at the center of the video game crash of 1983 due to the poor quality of its games at the time. It is a fascinating system in many ways that are still relevant today, especially when it comes to pushing hardware much farther than it was designed to go. [nicole] brings us a project that overcomes some of the limitations in its hardware to provide a more modern video output.

At the heart of the Atari is a custom chip called teh Television Interface Adapter (TIA) that generates the console’s video signal as well as handling controller information and a few other tasks. It was designed at a time where memory was expensive, and essentially trades programmer effort to reduce memory requirements. Interestingly, it separates luminance and chrominance information much like S-video does, so that’s where [nicole] focused their efforts. Thanks to some help from an adapter board, the video signals can be intercepted and reprocessed for the S-video standard instead of using RF modulation to send video data out, although this does involve some soldering and modifying of the original Atari hardware. In [nicole]’s case this was a little more involved due to the differences of the 2600jr compared to more standard versions of the console.

While S-video isn’t modern in the strictest sense, as a standard from 1987 it is a huge step forward compared to the available video output methods available in the 1970s when the 2600 was first produced. Plenty of older consoles and other hardware like VCRs and the like used S-video, so if you have a retro gaming setup complete with a CRT you might want to take a look at this 12-input A/V switch to keep everything managed.

A Fossil Wrist PDA running the Overbite Gopher browser

Mobile Gopher Client Brings Fossil Wrist PDA Online

Like many new technologies, smartwatches needed a few iterations before they became useful enough for the average person. Early examples were too clunky and limited to be of use to anyone but geeks who wanted to show off their “next big thing”. The 2005 Fossil Wrist PDA was a prime example: although impressively compact for its time, its limited battery life and poor feature set made it obsolete as soon as it was released. But since it ran on Palm OS, it offered plenty of opportunity for hacking: Palm expert [Cameron Kaiser] has upgraded his Wrist with internet access.

While Palm OS 4 natively supports TCP/IP networking, this component was deleted from the Wrist version to save memory. In any case, the only viable network interface would have been the USB port, which isn’t too convenient for a watch. Not to be deterred, [Cameron] worked out a way to add network support back into the Wrist: he used the IR port on a Palm m505 to send a copy of its own network drivers to the watch. This works because both devices run the same basic OS version on the same CPU type; the only drawback is that the network setup dialog doesn’t respond correctly to the Wrist’s different set of buttons. Continue reading “Mobile Gopher Client Brings Fossil Wrist PDA Online”

Kim-1: Memory Problem Resolved

At the very start of the personal computer revolution, there were relatively inexpensive boards with little more than a CPU, some memory, a display, and switches or a keypad. Some of these had expansion ports meant to allow you to build up, and some were just “trainers” to learn about computers. While you could argue that the Altair fell into this category, it had a case and a proper bus. The computers we are thinking about were usually just on a single board and — with luck — had an edge connector for expansion. Perhaps the most famous of these was the KIM-1 and [Old VCR] shows us how he brought one back to life.

These were highly popular mainly because of the low price of $245 back in 1976. For that price you got a calculator-style keyboard and LED display, 1K of RAM, and 2K of ROM. [Old VCR] has several and noticed that one was developing memory problems.

Continue reading “Kim-1: Memory Problem Resolved”

A Commodore 128 with a video capture device attached

Hacking The Commodore 128 To Capture Almost Real-Time Video

Although watching and editing videos may be among the primary tasks of many PCs today, it wasn’t that long ago that working with video required powerful processors and expensive video capture hardware. Even in the 1980s, home computer users were looking for ways to connect video sources to their Commodores and Ataris despite their hardware limitations. [Cameron Kaiser] has a mid-1980s consumer-grade video capture device, which he has managed to turn into an almost real-time video capture system.

A distorted video image on a C128's monitor
Allowing the graphics chip to interrupt the CPU mid-capture results in a severely distorted image

His work revolves around a device called “ComputerEyes”, a 1984-vintage hardware interface that made it possible to connect a composite video source to a home computer. The limitations of mid-1980s CPUs meant that it took around six seconds for the computer to do a quick scan of a single video frame, or a multiple of that if you wanted a higher-quality image. Another limitation, at least on Commodore machines, was that the screen had to be turned off during video capture – otherwise, the video chip would interrupt the CPU halfway through the process, causing it to lose its synchronization with the video source.

[Cameron] however, plugged his ComputerEyes into a Commodore 128. This machine, largely designed by Hackaday contributor [Bil Herd], has an unusual hardware architecture consisting of two different CPUs and, crucially, two separate video chips. The primary 8564 “VIC-II” graphics chip is used to keep compatibility with existing Commodore 64 programs, while the secondary 8563 “VDC” is mainly aimed at newer high-resolution text-based software. The VDC is also much more independent from the main system bus than the VIC-II, allowing it to display an image without disturbing the CPU.

More after the break.

Continue reading “Hacking The Commodore 128 To Capture Almost Real-Time Video”