The PalmPilot Returns, This Time In Your Browser

The PalmPilot doesn’t seem to get much retrocomputing love, but maybe it should. After all, it might not have been the very first handheld, but it was probably the most successful, and that ultimately led to the era of the smartphone. Whether you miss your old Palm applications, or never got to experience them the first time around, fear not. You can now relive them in all their glory in your browser thanks to the Internet Archive project.

There are over 500 applications and games all running in a browser-based emulator. Some of the programs don’t seem to work well, and some don’t make sense in the context of a virtual environment. But many work fine, and if you want the classic apps, just open up anything and press the home button. If you want a review of the Palm IIIe PDA from 1999, check out [VWestlife’s] video, below.

The Grafitti handwriting recognition system was state-of-the-art for the day. The key was the system could more easily recognize printing if it were mostly single strokes that always worked the same way. For example, the “A” had no crossbar and the “F” was missing the bottom horizontal line. As much as possible, you make letters with a single stroke and there was only one way to form each letter. Good times!

What was high tech back then you can now build out of spare parts. If you happen to have a Palm, you might consider giving it a much-needed backlight.

Continue reading “The PalmPilot Returns, This Time In Your Browser”

image of drfitwood on a beach

Ask Hackaday: Not Your Mother’s Feedback

Imagine you were walking down a beach, and you came across some driftwood resting against a pile of stones. You see it in the distance, and your brain has no trouble figuring out what you’re looking at. You see driftwood and rocks – you can clearly distinguish between the two objects without a second thought.

Think about the raw data entering the brain. The textures of the rocks and the driftwood are similar. The colors are similar. The irregular shapes are similar. Thus the raw data entering the brain’s V1 area for both objects must be similar as well. Now think about the borders that separate the pieces of driftwood from the edges of the rocks. From a raw data perspective, there is no border, and likewise no separation because the two objects are so similar.  But yet your brain can clearly see a rock and a piece of driftwood – two distinctly different objects. So how does the brain do this? How does it so easily differentiate between the two? If the raw data on either side of the border separating the wood and the rocks is the same, then there must be an outside influence determining where that border is. Jeff Hawkins believes this outside influence is a very special and most interesting type of feedback. Read on as we explain and attempt to implement this form of feedback in our hierarchical structure of invariant representations.

Continue reading “Ask Hackaday: Not Your Mother’s Feedback”

binary hierarchy

Ask Hackaday: Sequences Of Sequences

ask-hackaday-invariant-representations-featured-image

 

In a previous article, we talked about the idea of the invariant representation and theorized different ways of implementing such an idea in silicon. The hypothetical example of identifying a song without knowledge of pitch or form was used to help create a foundation to support the end goal – to identify real world objects and events without the need of predefined templates. Such a task is possible if one can separate the parts of real world data that changes from that which does not. By only looking at the parts of the data that doesn’t change, or are invariant, one can identify real world events with superior accuracy compared to a template based system.

Consider a friend’s face. Imagine they were sitting in front of you, and their face took up most of your visual space. Your brain identifies the face as your friend without trouble. Now imagine you were in a crowded nightclub, and you were looking for the same friend. You catch a glimpse of her from several yards away, and your brain ID’s the face without trouble. Almost as easily as it did when she was sitting in front of you.

I want you to think about the raw data coming off the eye and going into the brain during both scenarios. The two sets of data would be completely different. Yet your brain is able to find a commonality between the two events. How? It can do this because the data that makes up the memory of your friend’s face is stored in an invariant form. There is no template of your friend’s face in your brain. It only stores the parts that do not change – such as the distance between the eyes, the distance between the eye and the nose, or the ear and the mouth. The shape her hairline makes on her forehead. These types of data points do not change with distance, lighting conditions or other ‘noise’.

One can argue over the specifics of how the brain does this. True or not true, the idea of the invariant representation is a powerful one, and implementing such an idea in silicon is a worthy goal. Read on as we continue to explore this idea in ever deeper detail.

Continue reading “Ask Hackaday: Sequences Of Sequences”

Ask Hackaday: What Are Invariant Representations?

Your job is to make a circuit that will illuminate a light bulb when it hears the song “Mary Had a Little Lamb”. So you breadboard a mic, op amp, your favorite microcontroller (and an ADC if needed) and get to work. You will sample the incoming data and compare it to a known template. When you get a match, you light the light. The first step is to make the template. But what to make the template of?

“Hey boss, what style of the song do you want to trigger the light? Is it children singing, piano, what?”

Your boss responds:

“I want the light to shine whenever any version of the song occurs. It could be singing, keyboard, guitar, any musical instrument or voice in any key. And I want it to work even if there’s a lot of ambient noise in the background.”

Uh oh. Your job just got a lot harder. Is it even possible? How do you make templates of every possible version of the song? Stumped, you talk to your friend about your dilemma over lunch, who just so happens to be [Jeff Hawkins] – a guy whose already put a great deal of thought into this very problem.

“Well, the brain solves your puzzle easily.” [Hawkins] says coolly. “Your brain can recall the memory of that song no matter if it’s vocal, instrumental in any key or pitch. And it can pick it out from a lot of noise.”

“Yea, but how does it do that though!” you ask. “The pattern’s of electrical signals entering the brain have to be completely different for different versions of the song, just like the patterns from my ADC. How does the brain store the countless number of templates required to ID the song?”

“Well…” [Hawkins] chuckles. “The brain does not store templates like that”. The brain only remembers the parts of the song that doesn’t change, or are invariant. The brain forms what we call invariant representations of real world data.”

Eureka! Your riddle has been solved. You need to construct an algorithm that stores only the parts of the song that doesn’t change. These parts will be the same in all versions – vocal or instrumental in any key. It will be these invariant, unchanging parts of the song that you will look for to trigger the light. But how do you implement this in silicon?

Continue reading “Ask Hackaday: What Are Invariant Representations?”

Classic PDA Finds Second Life As A Network Touch Screen Display

pocketviewer

[Tomas Janco] had an old Casio Pocket Viewer PDA collecting dust. Rather than throw it away, He decided to re-purpose it as a display for time, weather, and the current status of his garage door.

The Casio Pocket Viewer was a competitor to the Palm Pilot. The two systems even shared the same LCD resolution – 160×160 monochrome. [Tomas’] particular model is an S660, sporting 6 megabytes of ram and an NEC V30MZ (Intel 8086 compatible) processor. Similar to Palm, Casio made an SDK freely available.

The SDK is still available from Casio, and [Tomas] was able to get it running on his PC. Development wasn’t without pitfalls though. The Pocket Viewer SDK was last updated in April of 2001. Software is written in C, but the then new C99 standard is not supported. The SDK does include a simulator and debugger, but it too is not as polished as todays systems – every simulator startup begins with setting the clock and calibrating the touch screen. Keep reading after the jump to learn about the rest of the hurdles he overcame to pull this one off.

Continue reading “Classic PDA Finds Second Life As A Network Touch Screen Display”