Retro-Inspired Computer Case Hosts Mechanical Keyboard

During the time in the 1980s when the personal computer was gaining steam as a household fixture, plenty of models shipped with the keyboard built in to the machine itself. This helped reduce costs, lower the physical footprint of the device, and arguably improved aesthetics. But as technology progressed, this type of design fell by the wayside as computers became more modular and configurable. That’s not to say there aren’t any benefits to building a computer like this, though. [jit] is here to show off this Amiga-inspired computer with its own modern built-in mechanical keyboard.

Like the Raspberry Pi 400 which is built into its own case, modern computers like this are extremely portable, relatively simple, and space-efficient. But [jit] did not like the uninspired design of the Pi so he was looking to make some improvements. Starting with the keyboard, it boasts a 60% size board with mechanical keys which are backlit by LEDs. Inside the machine is a Odroid XU4 which has a little bit more power (and is often easier to find) than a comparable Raspberry Pi. The case is 3D printed and includes ventilation and support for the addition of various cooling fans, I/O ports, status LEDs, and switches for the computer inside.

Additionally, some modification of the Odroid itself was needed in order to move the various switches to the case, and the build also includes a somewhat customized power supply internally as well. It’s a well-rounded build that captures the spirit of the old computer cases, but takes advantage of a lot of modern technology at the same time. If you want to go all-out with a build like this, though, take a look at this retro-inspired case (with keyboard included) that manages to get most of a Framework laptop inside.

Continue reading “Retro-Inspired Computer Case Hosts Mechanical Keyboard”

Ecological System Dynamics For Computing

Some of you may remember that the ship’s computer on Star Trek: Voyager contained bioneural gel packs. Researchers have taken us one step closer to a biocomputing future with a study on the potential of ecological systems for computing.

Neural networks are a big deal in the world of machine learning, and it turns out that ecological dynamics exhibit many of the same properties. Reservoir Computing (RC) is a special type of Recurrent Neural Network (RNN) that feeds inputs into a fixed-dynamics reservoir black box with training only occurring on the outputs, drastically reducing the computational requirements of the system. With some research now embodying these reservoirs into physical objects like robot arms, the researchers wanted to see if biological systems could be used as computing resources.

Using both simulated and real bacterial populations (Tetrahymena thermophila) to respond to temperature stimuli, the researchers showed that ecological system dynamics has the “necessary conditions for computing (e.g. synchronized dynamics in response to the same input sequences) and can make near-future predictions of empirical time series.” Performance is currently lower than other forms of RC, but the researchers believe this will open up an exciting new area of research.

If you’re interested in some other experiments in biocomputing, checkout these RNA-based logic gates, this DNA-based calculator, or this fourteen-legged state machine.

A Simple Guide To Bit Banged I2C On The 6502

We covered [Anders Nielsen]’s 65duino project a short while ago, and now he’s back with an update video showing some more details of bit-banging I2C using plain old 6502 assembly language.

Obviously, with such a simple system, there is no dedicated I2C interface hardware, so the programmer must take care of all the details of the I2C protocol in software, bit-banging it out to the peripheral and reading back the response one bit at a time.

The first detail to concern us will be the I2C addresses of the devices being connected to the bus and how low-level bit manipulation is used to turn the 7-bit I2C address into the byte being bit-banged. As [Anders] shows, setting a bit is simply a logical-OR operation, and resetting a bit is a simple logical-AND operation using the inversion (or one’s complement) bit to reset to form a bitmask. As many will already know, this process is necessary to code for a read or a write I2C operation. A further detail is that I2C uses an open-collector connection scheme, which means that no device on the bus may drive the bus to logical high; instead, they must release the drive by going to the high impedance state, and an external pull-up resistor will pull the bus high. The 6532 RIOT chip (used for I/O on the 65unio) does not have tristate control but instead uses a data direction register (DDR) to allow a pin to be an input. This will do the job just fine, albeit with slightly odd-looking code, until you know what’s going on.

From there, it’s a straightforward matter to write subroutines that generate the I2C start, stop, and NACK conditions that are required to write to the SSD1306-based OLED to get it to do something we can observe. From these basic roots, through higher-level subroutines, a complete OLED library in assembly can be constructed. We shall sit tight and await where [Anders] goes next with this!

We see I2C-connected things all the time, like this neat ATtiny85-based I2C peripheral, and whilst we’re talking about the SSD1306 OLED display controller, here’s a hack that shows just how much you can push your luck with the I2C spec and get some crazy frame rates.

Continue reading “A Simple Guide To Bit Banged I2C On The 6502”

Intel To Ship Quantum Chip

In a world of 32-bit and 64-bit processors, it might surprise you to learn that Intel is releasing a 12-bit chip. Oh, wait, we mean 12-qubit. That makes more sense. Code named Tunnel Falls, the chip uses tiny silicon spin quantum bits, which Intel says are more advantageous than other schemes for encoding qubits. There’s a video about the device below.

It is a “research chip” and will be available to universities that might not be able to produce their own hardware. You probably aren’t going to find them listed on your favorite online reseller. Besides, the chip isn’t going to be usable on a breadboard. It is still going to take a lot of support to get it running.

Intel claims the silicon qubit technology is a million times smaller than other qubit types. The size is on the order of a device transistor — 50 nanometers square — simplifying things and allowing denser devices. In silicon spin qubits, information resides in the up or down spin of a single electron.

Of course, even Intel isn’t suggesting that 12 qubits are enough for a game-changing quantum computer, but you do have to start somewhere. This chip may enable more researchers to test the technology and will undoubtedly help Intel accelerate its research to the next step.

There is a lot of talk that silicon is the way to go for scalable quantum computing. It makes you wonder if there’s anything silicon can’t do? You can access today’s limited quantum computers in the proverbial cloud.

Continue reading “Intel To Ship Quantum Chip”

Your IPhone Can’t Do What This WinCE Device Can!

Most of us probably now have a smartphone, an extremely capable pocket computer — even if sometimes its abilities are disguised a little by its manufacturer. There are many contenders to the crown of first smartphone, but in that discussion it’s often forgotten that the first generally available such devices weren’t phones at all, but PDAs, or Personal Digital Assistants. The fancier ones blurred the line between PDA and laptop and were the forerunner devices to netbooks, and it’s one of these that [Remy] is putting through its paces. He makes the bold claim that it can do things the iPhone can’t, and while the two devices are in no way comparable he’s right on one point. His HP Journada 720 can host a development environment, while the iPhone can’t.

The HP was something of a turn-of-the-millennium object of desire, being a palmtop computer with a half-decent keyboard a 640×240 pixel TFT display, and 32 MB of RAM alongside its 206 MHz Intel StrongARM CPU. Its Windows CE OS wasn’t quite the desktop Windows of the day, but it was close enough to be appealing for the ’90s exec who had everything. Astoundingly it has more than one Linux distro that can run on it with some level of modernity, which is where he’s able to make the claim about the iPhone being inferior.

We remember the Journada clamshell series from back in the day, though by our recollection the battery life would plummet if any attempt was made to use the PCMCIA slot. It was only one of several similar platforms offering a mini-laptop experience, and we feel it’s sad that there are so few similar machines today. Perhaps we’ll keep an eye out for one and relive the ’90s ourselves.

A Modular Analogue Computer

We are all used to modular construction in the analogue synth world, to the extent that there’s an accepted standard for it in EuroRack. But the same techniques are just as useful wherever else analogue circuits need to be configured on the fly, such as in an analogue computer. It’s something [Rainer Glaschick] has pursued, with his Flexible Analog Computer, an analogue computer made from a set of modules mounted on breadboard strips.

Standard modules are an adder and an integrator, with the adder also having inverter, comparator, and precision rectifier functions. The various functions can be easily configured by means of jumpers, and there are digital switches on board to enable or disable outputs and inputs. he’s set up a moon landing example to demonstrate the machine in practice.

We’re not going to pretend to be analogue computer experts here at Hackaday,but we naturally welcome any foray into analogue circuitry lest it become a lost art. If you’d like to experiment with analogue computing there are other projects out there to whet your appetite, and of course they don’t even need to be electronic.