There is more than one way to lead a successful life. Some people have all the opportunity in the world laid out before them, and it never does them any good. Others have little more than the determination and desire they’ve dredged up within themselves, and that grit turns out to be the abrasive that smooths the path ahead.
Ronald McNair succeeded despite poverty, racism, and an education system designed to keep Black people down. He became an accidental revolutionary at the age of nine, when he broke the color barrier in his small South Carolina town via the public library. This act of defiance in pursuit of education would set the course for his relatively short but full life, which culminated in his career as a Space Shuttle mission specialist.
Rule-Breaker with a Slide Rule
Ronald McNair was born October 21, 1950 in Lake City, South Carolina, the second of three sons, to Pearl and Carl McNair. His mother was a teacher, and encouraged his love of reading. Ronald’s father, Carl was an auto mechanic who never finished high school and always regretted it. Though the family was poor, Ron grew up surrounded by books, music, and support.
In 2020, our digital world and the software we use to create it are a towering structure, built upon countless layers of abstraction and building blocks — just think about all the translations and interactions that occur from loading a webpage. Whilst abstraction is undoubtedly a great thing, it only works if we’re building on solid ground; if the lower levels are stable and fast. What does that mean in practice? It means low-level, compiled languages, which can be heavily optimised and leveraged to make the most of computer hardware. One of the giants in this area was Frances Allen, who recently passed away in early August. Described by IBM as “a pioneer in compiler organization and optimization algorithms,” she made numerous significant contributions to the field. Continue reading “Frances Allen Optimised Your Code Without You Even Knowing”→
It’s true what they say — you never know what you can do until you try. Russell Kirsch, who developed the first digital image scanner and subsequently invented the pixel, was a firm believer in this axiom. And if Russell had never tried to get a picture of his three-month-old son into a computer back in 1957, you might be reading Hackaday in print right now. Russell’s work laid the foundation for the algorithms and storage methods that make digital imaging what it is today.
Russell A. Kirsch was born June 20, 1929 in New York City, the son of Russian and Hungarian immigrants. He got quite an education, beginning at Bronx High School of Science. Then he earned a bachelor’s of Electrical Engineering at NYU, a Master of Science from Harvard, and attended American University and MIT.
In 1951, Russell went to work for the National Bureau of Standards, now known as the National Institutes of Science and Technology (NIST). He spent nearly 50 years at NIST, and started out by working with one of the first programmable computers in America known as SEAC (Standards Eastern Automatic Computer). This room-sized computer built in 1950 was developed as an interim solution for the Census Bureau to do research (PDF).
Like the other computers of its time, SEAC spoke the language of punch cards, mercury memory, and wire storage. Russell Kirsch and his team were tasked with finding a way to feed pictorial data into the machine without any prior processing. Since the computer was supposed to be temporary, its use wasn’t as tightly controlled as other computers. Although it ran 24/7 and got plenty of use, SEAC was more accessible than other computers, which allowed time for bleeding edge experimentation. NIST ended up keeping SEAC around for the next thirteen years, until 1963.
The Original Pixel Pusher
The term ‘pixel’ is a shortened portmanteau of picture element. Technically speaking, pixels are the unit of length for digital imaging. Pixels are building blocks for anything that can be displayed on a computer screen, so they’re kind of the first addressable blinkenlights.
As the drum slowly rotated, a photo-multiplier moved back and forth, scanning the image through a square viewing hole in the wall of a box. The tube digitized the picture by transmitting ones and zeros to SEAC that described what it saw through the square viewing hole — 1 for white, and 0 for black. The digital image of Walden is 76 x 76 pixels, which was the maximum allowed by SEAC.
In in the video below, Russell discusses the idea and proves that variable pixels make a better image with more information than square pixels do, and with significantly fewer pixels overall. It takes some finagling, as pixel pairs of triangles and rectangles must be carefully chosen, rotated, and mixed together to best represent the image, but the image quality is definitely worth the effort. Following that is a video of Russell discussing SEAC’s hardware.
Russell retired from NIST in 2001 and moved to Portland, Oregon. As of 2012, he could be found in the occasional coffeehouse, discussing technology with anyone he could engage. Unfortunately, Russell developed Alzheimer’s and died from complications on August 11, 2020. He was 91 years old.
When you’re a kid, one of the surest signs of summer is hearing the happy sound of the ice cream truck crawling through the neighborhood. You don’t worry about how that magical truck is keeping the ice cream cold, only that it rolls down your street, and that the stars align and your parents give you money for a giant ice cream-cookie sandwich with the edge rolled in tiny chocolate chips.
In the early days of mobile refrigeration, ice cream trucks and other food delivery vehicles relied first on ice, and then dry ice to keep perishables cold. Someone eventually invented an electric cooling system, but those had to be recharged periodically at power stations. There was also a short-lived mechanical system, but it was highly susceptible to road vibrations.
Until Frederick McKinley Jones came along, mobile refrigeration was fledgling, and sources of perishable food were extremely localized and limited. In the early 1940s, Frederick patented the first practical automated refrigeration system for trucks, and it revolutionized the shipping and storage of food and medicine.
Every piece of technology starts with a vision, a vague notion of how a thing could or should be. The computer mouse is no different. In fact, the mouse was built to be an integral part of the future of personal computing — a shift away from punch cards and mystery toward a more accessible and user-friendly system of windowed data display, hyperlinks, videoconferencing, and more. And all of it would be commanded by a dot on the screen moving in sync with the operator’s intent, using a piece of hardware controlled by the hand.
The stuff of science fiction becomes fact anytime someone has the means to make it so. Often times the means includes another human being, a intellectual complement who can conjure the same rough vision and fill in the gaps. For Douglas Engelbart’s vision of the now-ubiquitous computer mouse, that person was William English.
William English was born January 27, 1929 in Lexington, Kentucky. His father was an electrical engineer and William followed this same path after graduating from a ranch-focused boarding school in Arizona. After a stint in the Navy, he took a position at Stanford Research Institute in California, where he met Douglas Engelbart.
Engelbart showed William his notes and drawings, and he built the input device that Englebart envisioned — one that could select characters and words on the screen and revolutionize text editing. The X/Y Position Indicator, soon and ever after called the mouse: a sort of rough-yet-sleek pinewood derby car of an input device headed into the future of personal computing.
William’s mouse was utilitarian: a wooden block with two perpendicular wheels on the bottom, and a pair of potentiometers inside to interpret the wheels’ X and Y positions. The analog inputs are converted to digital and represented on the screen. The first mouse had a single button, and the cord was designed to run out the bottom, not the top.
If you were to travel back in time to the turn of the previous century and try to convince the average person that the grains of sand on just about any beach would be the basis of an industry worth hundreds of billions of dollars within 100 years, they’d probably have thought you were crazy. Aside from being coarse, rough, and irritating, sand is everywhere, and convincing anyone of its value would be a hard sell, unless your interlocutor was a real estate visionary with an appreciation of the future value of seaside property and a lot of patience.
Fast forward to our time, and we all know the value of the material that comes from common quartz sand: silicon, specifically the ultra-purified crystals of silicon that end up as the wafers we depend on to build the circuitry of life. The trip from beach to chip foundry is a long and non-obvious one which would not have been possible without the insights of an undistinguished Polish student and one-time druggist who discovered the process that made the Information Age possible: Jan Czochralski.
The video game console is now a home entertainment hub that pulls in all forms of entertainment via an internet connection, but probably for most readers it was first experienced as an offline device that hooked up to the TV and for which new game software had to be bought as cartridges or for later models, discs. Stepping back through the history of gaming is an unbroken line to the 1970s, but which manufacturer had the first machine whose games could be purchased separately from the console? The answer is not that which first comes to mind, and the story behind its creation doesn’t contain the names you are familiar with today.
The Fairchild Channel F never managed to beat its rival, the Atari 2600, in the hearts of American youngsters so its creator Jerry Lawson isn’t a well-known figure mentioned in the same breath as Atari’s Nolan Bushnell or Apple’s two Steves, but without this now-forgotten console the history of gaming would have been considerably different.