The other issue is that solar cells have a guanteed life expectancy of about 25 years, with average efficiency losses of 0.5% per year. If replacement begins after 25 years, time is running out for all the panels that were installed during the early 2000s boom. The International Renewable Energy Agency (IREA) projects that by 2050, we’ll be looking at 78 million metric tons of bulky e-waste. The IREA also believe that we’ll be generating six million metric tons of new solar e-waste every year by then, too. Unfortunately, there are hardly any measures in place to recycle solar panels, at least in the US.
How are solar panels made, anyway? And why is it so hard to recycle them? Let’s shed some light on the subject.
You’d think that the 8086 microprocessor, a 40-year-old chip with a mere 29,000 transistors on board that kicked off the 16-bit PC revolution, would have no more tales left to tell. But as [Ken Shirriff] discovered, reverse engineering the chip from die photos reveals some hidden depths.
The focus of [Ken]’s exploration of the venerable chip is the charge pump, a circuit that he explains was used to provide a bias voltage across the substrate of the chip. Early chips generally took this -5 volt bias voltage from a pin, which meant designers had to provide a bipolar power supply. To reduce the engineering effort needed to incorporate the 8086 into designs, Intel opted for an on-board charge pump to generate the bias voltage. The circuit consists of a ring oscillator made from a trio of inverters, a pair of transistors, and some diodes to act as check valves. By alternately charging a capacitor and switching its polarity relative to the substrate, the needed -5 volt bias is created.
Given the circuit required, it was pretty easy for [Ken] to locate it on the die. The charge pump takes up a relatively huge amount of die space, which speaks to the engineering decisions Intel made when deciding to include it. [Ken] drills down to a very low level on the circuit, with fascinating details on how the MOSFETs were constructed, and why eight transistors were used instead of two diodes. As usual, his die photos are top quality, as are his explanations of what’s going on down inside the silicon.
If you were to travel back in time to the turn of the previous century and try to convince the average person that the grains of sand on just about any beach would be the basis of an industry worth hundreds of billions of dollars within 100 years, they’d probably have thought you were crazy. Aside from being coarse, rough, and irritating, sand is everywhere, and convincing anyone of its value would be a hard sell, unless your interlocutor was a real estate visionary with an appreciation of the future value of seaside property and a lot of patience.
Fast forward to our time, and we all know the value of the material that comes from common quartz sand: silicon, specifically the ultra-purified crystals of silicon that end up as the wafers we depend on to build the circuitry of life. The trip from beach to chip foundry is a long and non-obvious one which would not have been possible without the insights of an undistinguished Polish student and one-time druggist who discovered the process that made the Information Age possible: Jan Czochralski.
Chip decapping videos are a staple of the hacking world, and few things compare to the beauty of a silicon die stripped of its protective epoxy and photographed through a good microscope. But the process of actually opening that black resin treasure chest seems elusive, requiring as it does a witch’s brew of solvents and acids.
Or does it? As [Curious Marc] documents in the video below, a little heat and some finesse are all it takes, at least for some chips. The method is demonstrated by [Antoine Bercovici], a paleobotanist who sidelines as a collector of old chips. After removing chips from a PCB — he harvested these chips from an old PlayStation — he uses hot air to soften the epoxy, and then flexes the chip with a couple of pairs of pliers. It’s a bit brutal, but in most of the Sony chips he tried for the video, the epoxy broke cleanly over the die and formed a cleavage plane that allowed the die to be slipped out cleanly. The process is not unlike revealing fossils in sedimentary rocks, a process that he’s familiar with from his day job.
He does warn that certain manufacturers, like Motorola and National, use resins that tend to stick to the die more. It’s also clear that a hairdryer doesn’t deliver enough heat; when they switched to a hot air rework station, the success rate went way up.
The simplicity of this method should open the decapping hobby up to more people. Whether you just want to take pretty pictures or if reverse engineering is on your mind, put the white fuming nitric acid down and grab the heat gun instead.
If you’ve ever handled a chip with a really strange or highly inconvenient pinout and suspected that the reason had something to do with the inner workings, you may be interested to see [electronupdate]’s analysis of why the 4017 Decade Counter IC has such a weirdly nonintuitive pinout. It peeks into an IC design dating from the 1970s to see an example of the kind of design issues that can affect physical layout.
In the case of the 4017, once decapped and the inner workings exposed, things became more clear. Inside the chip are a bunch of flip-flops and NAND gates, laid out in a single layer. Some of the outputs (outputs 5 and 1 for example, physically on pins 1 and 2 respectively) share the same flip-flop.
The original design placed the elements in a way that made the most logical sense for routing and layout, which resulted in nice and tidy inner workings but an apparently illogical pinout. A lot of this is probably feeling familiar to anyone who has designed and routed a single-layer PCB, where being limited to one layer makes it important to get the most connections as directly near one another as possible.
Chip design has of course come a long way since the 70s, but there is forever some level of trade-off to be made between outward tidiness and inner design harmony. The next time you’re looking at a part with an apparently illogical pinout, there’s a fair chance it makes far more sense on the inside.
Given the accuracy of Moore’s Law to the development of integrated circuits over the years, one would think that our present day period is no different from the past decades in terms of computer architecture design. However, during the 2017 ACM Turing Award acceptance speech, John L. Hennessy and David A. Patterson described the present as the “golden age of computer architecture”.
Compared to the early days of MS-DOS, when designing user- and kernel-space interactions was still an experiment in the works, it certainly feels like we’re no longer in the infancy of the field. Yet, as the pressure mounts for companies to acquire more computational resources for running expensive machine learning algorithms on massive swaths of data, smart computer architecture design may be just what the industry needs.
Moore’s law predicts the doubling of transistors in an IC, it doesn’t predict the path that IC design will take. When that observation was made in 1965 it was difficult or even impossible to envision where we are today, with tools and processes so closely linked and widely available that the way we conceive processor design is itself multiplying.
After spending much of the 20th century languishing in development hell, electric cars have finally hit the roads in a big way. Automakers are working feverishly to improve range and recharge times to make vehicles more palatable to consumers.
With a strong base of sales and increased uncertainty about the future of fossil fuels, improvements are happening at a rapid pace. Oftentimes, change is gradual, but every so often, a brand new technology promises to bring a step change in performance. Silicon carbide (SiC) semiconductors are just such a technology, and have already begun to revolutionise the industry.
Mind The Bandgap
Traditionally, electric vehicles have relied on silicon power transistors in their construction. Having long been the most popular semiconductor material, new technological advances have opened it up to competition. Different semiconductor materials have varying properties that make them better suited for various applications, with silicon carbide being particularly attractive for high-power applications. It all comes down to the bandgap.
Electrons in a semiconductor can sit in one of two energy bands – the valence band, or the conducting band. To jump from the valence band to the conducting band, the electron needs to reach the energy level of the conducting band, jumping the band gap where no electrons can exist. In silicon, the bandgap is around 1-1.5 electron volts (eV), while in silicon carbide, the band gap of the material is on the order of 2.3-3.3 eV. This higher band gap makes the breakdown voltage of silicon carbide parts far higher, as a far stronger electric field is required to overcome the gap. Many contemporary electric cars operate with 400 V batteries, with Porsche equipping their Taycan with an 800 V system. The naturally high breakdown voltage of silicon carbide makes it highly suited to work in these applications.