In 2003, nothing could stop AMD. This was a company that moved from a semiconductor company based around second-sourcing Intel designs in the 1980s to a Fortune 500 company a mere fifteen years later. AMD was on fire, and with almost a 50% market share of desktop CPUs, it was a true challenger to Intel’s throne.
AMD began its corporate history like dozens of other semiconductor companies: second sourcing dozens of other designs from dozens of other companies. The first AMD chip, sold in 1970, was just a four-bit shift register. From there, AMD began producing 1024-bit static RAMs, ever more complex integrated circuits, and in 1974 released the Am9080, a reverse-engineered version of the Intel 8080.
AMD had the beginnings of something great. The company was founded by [Jerry Sanders], electrical engineer at Fairchild Semiconductor. At the time [Sanders] left Fairchild in 1969, [Gordon Moore] and [Robert Noyce], also former Fairchild employees, had formed Intel a year before.
While AMD and Intel shared a common heritage, history bears that only one company would become the king of semiconductors. Twenty years after these companies were founded they would find themselves in a bitter rivalry, and thirty years after their beginnings, they would each see their fortunes change. For a short time, AMD would overtake Intel as the king of CPUs, only to stumble again and again to a market share of ten to twenty percent. It only takes excellent engineering to succeed, but how did AMD fail? The answer is Intel. Through illegal practices and ethically questionable engineering decisions, Intel would succeed to be the current leader of the semiconductor world.
With the Raspberry Pi and sever other ARM dev boards seeing their time in the lime light, it’s no surprise other chip manufacturers would want to get in on the action. AMD is releasing a very tiny x86 dev board called the Gizmo, a four-inch square board that shrinks a desktop computer down to the palm of your hand.
The Gizmo is powered by a dual-core x86 Brazos CPU running at 1 GHz with an included Radeon HD 6250 graphics engine. Also on the board is 1GB of DDR3 RAM, a SATA, Ethernet, USB, VGA, Audio, PCI and PCIe ports, and a ton of GPIO pins that include ADCs and DACs. All this in a four-inch square package that boasts about twice the performance of a Raspberry Pi.
While the price of the Gizmo – $200 for an explorer kit – will probably preclude it from being as popular as a Raspberry Pi or other ARM board, sometimes you just need an x86 platform to do the job. With the powerful graphics potential of the Gizmo, we could easily see this board being used in a few computer vision or autonomous robot builds.
[Gpuhackr] chose his username to explain exactly how he spends his time. For instance, here he’s using an STM32 Discovery board to drive an AMD Radeon HD 2400 graphics card. The ARM microcontroller isn’t actually using the PCIe interface on the card. Instead, [Gpuhackr] has patched into the debugging interface built into the card itself. This isn’t quite as straight forward as it sounds, but if you do the wiring carefully it’s a pretty intersting way to connect an ARM to an LCD monitor.
This project would be almost impossible if it weren’t for the open source code which AMD has released. This lets him implement the card’s 3D rendering features. The demo directly programs the UVD Xtensa CPU which is on the video card. It draws a cube with color gradients on each side. The cube spins while the debug information is overlaid on the screen. In this case the ARM chip/board is really being used as a programmer to upload some custom firmware. But we think a real code-ninja could implement a communications protocol to open up a simple way to drive the card in real-time.
[Hamster] wanted to take a look at division operations when the chip you’re using doesn’t have a divide instruction. He makes the point that the divide instruction takes a lot of space on the die, and that’s why it’s sometimes excluded from a chip’s instruction set. For instance, he tells us the ARM processor used on the Raspberry Pi doesn’t have a divide instruction.
Without hardware division you’re left to implement a binary division algorithm. Eventually [Hamster] plans to do this in an FPGA, but started researching the project by comparing division algorithms in C on an AMD processor.
His test uses all 16-bit possibilities for dividend and divisor. He was shocked to find that binary division doesn’t take much longer than using the hardware instruction for the same tests. A bit of poking around in his code and he manages to beat the AMD hardware divide instruciton by 175%. When testing with an Intel chip the hardware beats his code by about 62%.
He’s got some theories on why he’s seeing these performance differences which we’ll let you check out on your own.
Looks like some hardware enthusiasts have worked out a method to enable debug mode within AMD processors. The original site isn’t loading for us, but the text has been mirrored in this comment. Getting the chip into debug mode requires access passwords on four control registers. We’ve read through the writeup and it means very little to us but we didn’t pull out a datasheet to help make sense of the registers being manipulated. It shouldn’t be hard to find an old AMD system to try this out on. We’d love to hear about anything you do with this debug system.
Recently, research students at Georgia Tech released a report outlining the dangers that GPUs pose to the current state of password security. There are a number of ways to crack a password, all with their different pros and cons, but when it comes down to it, the limiting factor in all of these methods is processing complexity. The more operations that need to be run, the longer it takes, and the less useful each tool is for cracking passwords. In the past, most recommendations for password security revolved around making sure your password wasn’t something predictable, such as “password” or your birthday. With today’s (and tomorrows) GPUs, this may no longer be enough.
With most of us utilizing at least two monitors these days in our day to day operations, six monitors, while an awesome thought, might seem a little too excessive. After all, do we really have space for multiple video cards?
AMD has a new setup in their testing lab that is running six Dell 30inch displays at 7680×3200 through a video card holding six DisplayPort connectors.
Maximum PC has the scoop on the setup, and they say that this single GPU will be coming out on AMD’s DirectX 11 capable chips. Details are slim with the amount of video RAM, speeds and cost not known at present.
Think of the possibilities! Trade shows are one thing, but how about a video wall at home for gaming and movies? How would you use the six monitors shown above? Be sure to leave your ideas in the comments.