AMD is one of the oldest designers of large scale microprocessors and has been the subject of polarizing debate among technology enthusiasts for nearly 50 years. Its story makes for a thrilling tale -- filled with heroic successes, foolhardy errors, and a close shave with rack and ruin. Where other semiconductor firms have come and gone, AMD has weathered many storms and fought numerous battles, in boardrooms, courts, and stores.
In this feature we'll revisit the company's past, examine the twists and turns in the path to the present, and wonder at what lies ahead for this Silicon Valley veteran.
The rise to fame and fortune
To begin our story, we need to roll back the years and head for America and the late 1950s. Thriving after the hard years of World War II, this was the time and place to be if you wanted experience the forefront of technological innovation.
Companies such as Bell Laboratories, Texas Instruments, and Fairchild Semiconductor employed the very best engineers, and churned out numerous firsts: the bipolar junction transistor, the integrated circuit, and the MOSFET (metal oxide semiconductor field effect transistor).
These young technicians wanted to research and develop ever more exciting products, but with cautious senior managers mindful of the times when the world was fearful and unstable, frustration amongst the engineers build a desire to strike out alone.
And so, in 1968, two employees of Fairchild Semiconductor, Robert Noyce and Gordon Moore, left the company and forged their own path. N M Electronics opened its doors in that summer, to be renamed just weeks later as Integrated Electronics -- Intel, for short.
Others followed suit and less than a year later, another 8 people left and together they set up their own electronics design and manufacturing company: Advanced Micro Devices (AMD, naturally).
The group was headed by Jerry Sanders, Fairchild's former director of marketing, They began by redesigning parts from Fairchild and National Semiconductor rather than trying to compete directly with the likes of Intel, Motorola, and IBM (who spent significant sums of money on research and development of new integrated circuits).
From these humble beginnings, and a quick move from Santa Clara to Sunnyvale (Silicon Valley in California), AMD offered products that boasted increased efficiency, stress tolerances, and speed within a few months. These microchips were designed to comply with US military quality standards, which proved a considerable advantage in the still-young computer industry, where reliability and production consistency varied greatly.
By the time Intel released their first 8-bit microprocessor (the 8008) in 1974, AMD was a public company with a portfolio of over 200 products -- a quarter of which were their own designs, including RAM chips, logic counters, and bit shifters. The following year saw a raft of new models: their own Am2900 integrated circuit (IC) family and the 2 MHz 8-bit Am9080, a reverse-engineered copy of Intel's successor to the 8008. The former was a collection of components that are now fully integrated in CPUs and GPUs, but 35 years ago, arithmetic logic units and memory controllers were all separate chips.
The blatant plagiarism of Intel's design might seem to be somewhat shocking by today's standards, but it was par for the course in the fledgling days of microchips.
The blatant plagiarism of Intel's design might seem to be somewhat shocking by today's standards, but it was par for the course in the fledgling days of microchips. The CPU clone was eventually renamed as the 8080A, after AMD and Intel signed a cross-licensing agreement in 1976. You'd imagine this would cost a pretty penny or two, but it was just $325,000 ($1.65 million in today's dollars).
The deal allowed AMD and Intel to flood the market with ridiculously profitable chips, retailing at just over $350 or twice that for 'military' purchases. The 8085 (3 MHz) processor followed in 1977, and was soon joined by the 8086 (8 MHz). Design and manufacturing improvements led to the 8088 (5 to 10 MHz) appearing in 1979, the same year that also saw production begin at AMD's Austin, Texas facility.
When IBM began moving from mainframe systems into so-called personal computers (PCs) in 1982, the outfit decided to outsource parts rather than develop processors in-house. Intel's 8086, the first ever x86 processor, was chosen with the express stipulation that AMD acted as a secondary source to guarantee a constant supply for IBM's PC/AT.
A contract between AMD and Intel was signed in February of that year, with the former producing 8086, 8088, 80186, and 80188 processors -- not just for IBM, but for the many IBM clones that proliferated (Compaq being just one of them). AMD also started manufacturing the 16-bit Intel 80286, badged as the Am286, towards the end of 1982.
This was to become the first truly significant desktop PC processor, and while Intel's models generally ranged from 6 to 10 MHz, AMD's started at 8 MHz and went as high as 20 MHz. This undoubtedly marked the start of the battle for CPU dominance between the two Silicon Valley powerhouses; what Intel designed, AMD simply tried to make better.
This period represented a huge growth of the fledgling PC market, and noting that AMD had offered the Am286 with a significant speed boost over the 80286, Intel attempted to stop AMD in its tracks. This was done by excluding them from gaining a licence for the next generation 386 processors.
AMD sued, but arbitration took four and a half years to complete, and while the judgment found that Intel was not obligated to transfer every new product to AMD, it was determined that the larger chipmaker had breached an implied covenant of good faith.
Intel's licence denial occurred during a critical period, right as IBM PC's market was ballooning from 55% to 84%. Left without access to new processor specifications, AMD took over five years to reverse-engineer the 80386 into the Am386. Once completed, it proved once more to be more than a match for Intel's model. Where the original 386 debuted at just 12 MHz in 1985, and later managed to reach 33 MHz, the top-end version of the Am386DX launched in 1989 at 40 MHz.
The Am386's success was followed by the release of 1993's highly competitive 40 MHz Am486, which offered roughly 20% more performance than Intel's 33 MHz i486 for the same price. This was to be replicated throughout the entire 486 line up, and while Intel's 486DX topped out at 100 MHz, AMD offered (somewhat predictably at this stage) a snappier 120 MHz option. To better illustrate AMD's good fortune in this period, the company's revenue doubled from just over $1 billion in 1990 to well over $2 billion in 1994.
In 1995, AMD introduced the Am5x86 processor as a successor to the 486, offering it as a direct upgrade for older computers.
In 1995, AMD introduced the Am5x86 processor as a successor to the 486, offering it as a direct upgrade for older computers. The Am5x86 P75+ boasted a 150 Mhz frequency, with the 'P75' referencing performance that was similar to Intel's Pentium 75. The '+' signified that the AMD chip was slightly faster at integer math than the competition.
To counter this, Intel altered its naming conventions to distance itself from products by its rival and other vendors. The Am5x86 generated significant revenue for AMD, both from new sales and for upgrades from 486 machines. As with the Am286, 386 and 486, AMD continued to extend the market scope of the parts by offering them as embedded solutions.
March 1996 saw the introduction of its first processor, developed entirely by AMD's own engineers: the 5k86, later renamed to the K5. The chip was designed to compete with the Intel Pentium and Cyrix 6x86, and a strong execution of the project was pivotal to AMD -- the chip was expected to have a much more powerful floating point unit than Cyrix's and about equal to the Pentium 100, while the integer performance targeted the Pentium 200.
Ultimately, it was a missed opportunity, as the project was dogged with design and manufacturing issues. These resulted in the CPU not meeting frequency and performance goals, and it arrived late to market, causing it to suffer poor sales.
By this time, AMD had spent $857 million in stock on NexGen, a small fabless chip (design-only) company whose processors were made by IBM. AMD's K5 and the developmental K6 had scaling issues at higher clock speeds (~150 MHz and above) while NexGen's Nx686 had already demonstrated a 180 MHz core speed. After the buyout, the Nx686 became AMD’s K6 and the development of the original chip was consigned to the scrapyard.
The K6-2 introduced AMD's 3DNow! SIMD (single instruction, multiple data) instruction set.
AMD's rise reflected Intel's decline, from the early beginnings of the K6 architecture, which was pitted against Intel's Pentium, Pentium II and (largely rebadged) Pentium III. The K6 produced a quickening of AMD's success, owing its existence and capabilities to an ex-Intel employee, Vinod Dham (a.k.a. the "Father of Pentium"), who left Intel in 1995 to work at NexGen.
When the K6 hit shelves in 1997, it represented a viable alternative to the Pentium MMX. The K6 went from strength to strength -- from a 233 MHz speed in the initial stepping, to 300 MHz for the "Little Foot" revision in January 1998, 350 MHz in the "Chomper" K6-2 of May 1998, and an astonishing 550 MHz in September 1998 with the "Chomper Extended" revision.
The K6-2 introduced AMD's 3DNow! SIMD (single instruction, multiple data) instruction set. Essentially the same as Intel's SSE, it offered an easier route accessing the CPU's floating point capabilities; the downside to this being that programmers needed to incorporate the new instruction in any new code, in addition to patches and compilers needing to be rewritten to utilize the feature.
Like the initial K6, the K6-2 represented much better value than the competition, often costing half as much as Intel's Pentium chips. The final iteration of the K6, the K6-III, was a more complicated CPU, and the transistor count now stood at 21.4 million -- up from 8.8 million in the first K6, and 9.4 million for the K6-II.
It incorporated AMD's PowerNow!, which dynamically altered clock speeds according to workload. With clock speeds eventually reaching 570MHz, the K6-III was fairly expensive to produce and had a relatively short life span cut short by the arrival of the K7 which was better suited to compete with the Pentium III and beyond.
1999 was the zenith of AMD's golden age -- the arrival of the K7 processor, branded Athlon, showed that they were truly no longer the cheap, copycat option.
Starting at 500 MHz, Athlon CPUs utilized the new Slot A (EV6) and a new internal system bus licensed from DEC that operated at 200MHz, eclipsing the 133MHz Intel offered at the time. June 2000 brought the Athlon Thunderbird, a CPU cherished by many for its overclockability, which incorporated DDR RAM support and a full speed Level 2 on-die cache.
Thunderbird and its successors (Palomino, Thoroughbred, Barton, and Thorton), battled Intel's Pentium 4 throughout the first five years of the millennium, usually at a lower price point but always with better performance. Athlon was upgraded in September 2003 with the K8 (codenamed ClawHammer), better known as the Athlon 64, because it added a 64-bit extension to the x86 instruction set.
This episode is usually cited as AMD's defining moment. While it was was surging ahead, the MHz-at-any-cost approach of Intel's Netburst architecture was being exposed as a classic example of a developmental dead end.
Revenue and operating income were both excellent for such a relatively small company. While not Intel levels of income, AMD was flush with success and hungry for more. But when you're at the very peak of the tallest of mountains, it takes every ounce of effort to stay there -- otherwise, there's only way to go.
There is no single event responsible for AMD tumbling from its lofty position. A global economy crisis, internal mismanagement, poor financial predictions, a victim of its own success, the fortunes and misdeeds of Intel -- these all played a part, in some way or another.
But let's start seeing how matters were in early 2006. The CPU market was replete with offerings from both AMD and Intel, but the former had the likes of the exceptional K8-based Athlon 64 FX series. The FX-60 was a dual-core 2.6 GHz, whereas the FX-57 was single core, but ran at 2.8 GHz.
Both were head-and-shoulders above anything else, as shown by reviews at the time. They were hugely expensive, with the FX-60 retailing at over $1,000, but so was Intel's creme-de-la-creme, the 3.46 GHz Pentium Extreme Edition 955. AMD seemed to have the upper hand in the workstation/server market as well, with Opteron chips outperforming Intel's Xeon processors.
The problem for Intel was their Netburst architecture -- the ultra-deep pipeline structure required very high clock speeds to be competitive, which in turn increased power consumption and heat output. The design had reached its limits and was no longer up to scratch, so Intel ditched its development and turned to their older Pentium Pro/Pentium M CPU architecture to build a successor for the Pentium 4.
The initiative first produced the Yonah design for mobile platforms and then the dual-core Conroe architecture for desktops, in August 2006. Such was Intel's need to save face that they relegated the Pentium name to low-end budget models and replaced it with Core -- 13 years of brand dominance swept away in an instant.
The move to a low-power, high-throughput chip design wound up being ideally suited to a multitude of evolving markets and almost overnight, Intel took the performance crown in the mainstream and enthusiast sectors. By the end of 2006, AMD had been firmly pushed from the CPU summit, but it was a disastrous managerial decision that pushed them right down the slope.
Three days before Intel launched the Core 2 Duo, AMD made public a move that had been fully approved by then-CEO Hector Ruiz (Sanders had retired 4 years earlier). On July 24 2006, AMD announced that it intended to acquire the graphics card manufacturer ATI Technologies, in a deal worth $5.4 billion (comprising $4.3 billion in cash and loans, and $1.1 billion raised from 58 million shares). The deal was a huge financial gamble, representing 50% of AMD's market capitalization at the time, and while the purchase made sense, the price absolutely did not.
Imageon, the handheld graphics division of ATI, was sold to Qualcomm in a paltry $65 million deal. That division is now named Adreno, an anagram of "Radeon" and an integral component of the Snapdragon SoC
ATI was grossly overvalued, as it wasn't (nor was Nvidia) pulling in anything close to that kind of revenue. ATI had no manufacturing sites, either -- it's worth was almost entirely based on intellectual property.
AMD eventually acknowledged this mistake when it absorbed $2.65 billion in write-downs due to overestimating ATI's goodwill valuation.
To compound the lack of management foresight, Imageon, the handheld graphics division of ATI, was sold to Qualcomm in a paltry $65 million deal.That division is now named Adreno, an anagram of "Radeon" and an integral component of the Snapdragon SoC (!).
Xilleon, a 32-bit SoC for Digital TV and TV cable boxes, was shipped off to Broadcom for $192.8 million.
In addition to burning money, AMD's eventual response to Intel's refreshed architecture was distinctly underwhelming. Two weeks after Core 2's release, AMD's President and COO, Dirk Meyer, announced the finalization of AMD's new K10 Barcelona processor. This would be their decisive move in the server market, as it was a fully-fledged quad core CPU, whereas at the time, Intel only produced dual core Xeon chips.
The new Opteron chip appeared in September 2007, to much fanfare, but instead of stealing Intel's thunder, the party officially halted with the discovery of a bug that in rare circumstances could result in lockups when involving nested cache writes. Rare or not, the TLB bug put a stop to AMD's K10 production; in the meantime, a BIOS patch that cured the problem on outgoing processors, would do so at the loss of roughly 10% performance. By the time the revised 'B3 stepping' CPUs shipped 6 months later, the damage had been done, both for sales and reputation.
A year later, near the end of 2007, AMD brought the quad-core K10 design to the desktop market. By then, Intel was forging ahead and had released the now-famous Core 2 Quad Q6600. On paper, the K10 was the superior design -- all four cores were in the same die, unlike the Q6600 which used two separate dies on the same package. However, AMD was struggling to hit expected clock speeds, and the best version of the new CPU was just 2.3 GHz. That was slower, albeit by 100 MHz, than the Q6600, but it was also a little more expensive.
But the most puzzling aspect of it all was AMD's decision to come out with a new model name: Phenom. Intel switched to Core because Pentium had become synonymous with excessive price and power, and having relatively poor performance. On the other hand, Athlon was a name that computing enthusiasts knew all too well and it had the speed to match its reputation. The first version of Phenom wasn't actually bad -- it just wasn't as good as the Core 2 Quad Q6600, a product that was already readily available, plus Intel had faster offerings on the market, too.
Bizarrely, AMD seemed to make a conscious effort to abstain from advertising. They also had zero presence on the software side of the business; a very curious way to run a business, let alone one fighting in the semiconductor trade. But no review of this era in AMD's history would be complete without taking into consideration Intel's anti-competitive deeds. At this juncture, AMD was not only fighting Intel's chips, but also the company's monopolistic activities, which included paying OEMs large sums of money -- billions in total -- to actively keep AMD CPUs out of new computers.
In the first quarter of 2007, Intel paid Dell $723 million to remain the sole provider of its processors and chipsets -- accounting for 76% of the company's total operating income of $949 million. AMD would later win a $1.25 billion settlement in the matter, surprisingly low on the surface, but probably exacerbated by the fact that at the time of Intel's shenanigans, AMD itself couldn't actually supply enough CPUs to its existing customers.
Not that Intel needed to do any of this. Unlike AMD, they had rigid long-term goal setting, as well as greater product and IP diversity. They also had cash reserves like no one else: by the end of the first decade in the new millenium, Intel was pulling in over $40 billion in revenue and $15 billion in operating income. This provided huge budgets for marketing, research and software development, as well as foundries uniquely tailored to its own products and timetable. Those factors alone ensured AMD struggled for market share.
A multi-billion dollar overpayment for ATI and attendant loan interest, a disappointing successor to the K8, and problematic chips arriving late to market, were all bitter pills to swallow. But matters were about to get worse.
One step forward, one sideways, any number back
By 2010, the global economy was struggling to rebound from the financial crisis of 2008. AMD had ejected its flash memory section a few years earlier, along with all its chip making foundries -- they ultimately became GlobalFoundries, which AMD still uses for some of its products. Roughly 10% of its workforce had been dropped, and all together the savings and cash injection meant that AMD could knuckle down and focus entirely on processor design.
Rather than improving the K10 design, AMD started afresh with a new structure, and towards the end of 2011, the Bulldozer architecture was launched. Where K8 and K10 were true multicore, simultaneous multithreaded (SMT) processors, the new layout was classed as being 'clustered multithreading.'
AMD took a shared, modular approach with Bulldozer -- each cluster (or module) contained two integer processing cores, but they weren't totally independent. They shared the L1 instruction and L2 data caches, the fetch/decode, and the floating point unit. AMD even went as far as to drop the Phenom name and hark back to their glory days of the Athlon FX, by simply naming the first Bulldozer CPUs as AMD FX.
The idea behind all of these changes was to reduce the overall size of the chips and make them more power efficient. Smaller dies would improve fabrication yields, leading to better margins, and the increase in efficiency would help to boost clock speeds. The scalable design would also make it suitable for a wider range of markets.
The best model in the October 2011 launch, the FX-8510, sported 4 clusters but was marketed as an 8 core, 8 thread CPU. By this era, processors had multiple clock speeds, and the FX-8150 base frequency was 3.6 GHz, with a turbo clock of 4.2 GHz. However, the chip was 315 square mm in size and had a peak power consumption of over 125 W. Intel had already released the Core i7-2600K: it was a traditional 4 core, 8 thread CPU, running at up to 3.8 GHz. It was significantly smaller than the new AMD chip, at 216 square mm, and used 30 W less power.
On paper, the new FX should have dominated, but its performance was somewhat underwhelming -- at times, the ability to handle lots of threads would shine through, but single threaded performance was often no better than the Phenom range it was set to replace, despite the superior clock speeds.
Having invested millions of dollars into Bulldozer's R&D, AMD was certainly not going to abandon the design and the purchase of ATI was now starting to bear fruit. In the previous decade, AMD's first foray into a combined CPU and GPU package, called Fusion, was late to market and disappointingly weak.
But the project provided AMD with the means with which to tackle another markets. Earlier in 2011, another new architecture had been released, called Bobcat.
Aimed at low power applications, such as embedded systems, tablets, and notebooks; it was also the polar opposite design to Bulldozer: just a handful of pipelines and nothing much else. Bobcat received a much needed update a few years later, into the Jaguar architecture, and was selected by Microsoft and Sony to power the Xbox One and PlayStation 4 in 2013.
Although the profit margins would be relatively slim as consoles are typically built down to the lowest possible price, both platforms sold in the millions and this highlighted AMD's ability to create custom SoCs.
AMD's Bobcat received an update into the Jaguar architecture, and was selected by Microsoft and Sony to power the Xbox One and PlayStation 4 in 2013.
AMD continued revising the Bulldozer design over the years -- Piledriver came first and gave us the FX-9550 (a 220 W, 5 GHz monstrosity), but Steamroller and the final version, Excavator (launched in 2011, with products using it 4 years later), were more focused on reducing the power consumption, rather than offering anything particularly new.
By then, the naming structure for CPUs had become confusing, to say the least. Phenom was long resigned to the history books, and FX was having a somewhat poor reputation. AMD abandoned that nomenclature and just labelled their Excavator desktop CPUs as the A-series.
The graphics section of the company, fielding the Radeon products, had similarly mixed fortunes. AMD retained the ATI brand name until 2010, swapping it to their own. They also completely rewrote the GPU architecture created by ATI in late 2011, with the release of Graphics Core Next (GCN). This design would last for nearly 8 years, finding its way into consoles, desktop PCs, workstations and servers; it's still in use today as the integrated GPU in AMD's so-called APU processors.
GCN processors grew to have immense compute performance, but the structure wasn't the easiest to get the best out of it. The most powerful version AMD ever made, the Vega 20 GPU in the Radeon VII, boasted 13.4 TFLOPs of processing power and 1024 GB/s of bandwidth -- but in games, it just couldn't reach the same heights as the best from Nvidia.
Radeon products often came with a reputation for being hot, noisy, and very power hungry. The initial iteration of GCN, powering the HD 7970, required well over 200 W of power at full load -- but it was manufactured on a relatively large process node, TSMC's 28nm. By the time GCN had reached full maturity, in the form of the Vega 10, the chips were being made by GlobalFoundries on their 14nm node, but energy requirements were no better with the likes of the Radeon RX Vega 64 consuming a maximum of nearly 300 W.
While AMD had decent product selection, they just weren't as good as they should have been, and they struggled to earn enough money.
|Financial Year||Revenue ($ billion)||Gross Margin||Operating Income ($ million)||Net Income ($ million)|
By the end of 2016, the company's balance sheet had taken a loss for 4 consecutive years (2012's financials were battered by a $700 million GlobalFoundries final write off). Debt was still high, even with the sale of its foundries and other branches, and not even the success of the system package in the Xbox and PlayStation provided enough help.
At face value, AMD looked to be in deep trouble.
New stars a-ryze
With nothing left to sell and no sign of any large investments coming to save them, AMD could only do one thing: double-down and restructure. In 2012, they picked up two people who would come to play vital roles in the revival of the semiconductor company.
Jim Keller, the former lead architect for the K8 range, had returned after a 13 year absence and set about heading up two projects: one an ARM-based design for the server markets, the other a standard x86 architecture, with Mike Clark (lead designer of Bulldozer) being the chief architect.
Joining him was Lisa Su, who had been Senior Vice President and General Manager at Freescale Semiconductors. She took up the same position at AMD and is generally credited, along with then-President Rory Read, as being behind the company's move into markets beyond the PC, especially consoles.
Two years after Keller's restoration in AMD's R&D section, CEO Rory Read stepped down and the SVP/GM moved up. With a doctorate in electronic engineering from MIT and having conducted research into SOI (silicon-on-insulator) MOSFETS, Su had the academic background and the industrial experience needed to return AMD to its glory days. But nothing happens overnight in the world of large scale processors -- chip designs take several years, at best, before they are ready for market. AMD would have to ride the storm until such plans could come to fruition.
While AMD continued to struggle, Intel went from strength to strength. The Core architecture and fabrication process nodes had matured nicely, and at the end of 2016, they posted a revenue of almost $60 billion. For a number of years, Intel had been following a 'tick-tock' approach to processor development: a 'tick' would be a new architecture, whereas a 'tock' would be a process refinement, typically in the form of a smaller node.
However, not all was well behind the scenes, despite the huge profits and near-total market dominance. In 2012, Intel expected to be releasing CPUs on a cutting-edge 10nm node within 3 years. That particular tock never happened -- indeed, the clock never really ticked, either. Their first 14nm CPU, using the Broadwell architecture, appeared in 2015 and the node and fundamental design remained in place for half a decade.
The engineers at the foundries repeatedly hit yield issues with 10nm, forcing Intel to refine the older process and architecture each year. Clock speeds and power consumption climbed ever higher, but no new designs were forthcoming; an echo, perhaps, of their Netburst days. PC customers were left with frustrating choices: choose something from the powerful Core line, but pay a hefty price, or choose the weaker and cheaper FX/A-series.
But AMD had been quietly building a winning set of cards and played their hand in February 2016, at the annual E3 event. Using the eagerly awaited Doom reboot as the announcement platform, the completely new Zen architecture was revealed to the public.
Very little was said about the fresh design besides phrases such as 'simultaneous multithreading', 'high bandwidth cache,' and 'energy efficient finFET design.' More details were given during Computex 2016, including a target of a 40% improvement over the Excavator architecture.
To say this was ambitious would be an understatement
To say this was ambitious would be an understatement -- especially in light of the fact that AMD had delivered modest 10% increases with each revision of the Bulldozer design, at best.
It would take them another 12 months before the chip actually appeared, but when it did, AMD's long-stewing plan was finally clear.
Any new hardware design needs the right software to sell it, but multi-threaded CPUs were facing an uphill battle. Despite consoles sporting 8-core CPUs, most games were still perfectly fine with just 4. The main reasons were Intel's market dominance and the design of AMD's chip in the Xbox One and PlayStation 4. The former had released their first 6-core CPU back in 2010, but it was hugely expensive (nearly $1,100). Others rapidly appeared, but it would be another seven years before Intel offered a truly affordable hexa-core processor, the Core i5-8400, at under $200.
The issue with console processors was that the CPU layout consisted of two 4-core CPUs in the same die, and there was high latency between the two sections of the chip. So game developers tended to keep the engine's threads located on one of the sections, and only use the other for general background processes. Only in the world of workstations and servers there was a need for seriously multi-threaded CPUs -- until AMD decided otherwise.
In March 2017, general desktop users could upgrade their systems with one of two 8-core,16-thread CPUs. A completely new architecture clearly deserved a new name, and AMD cast off Phenom and FX, to give us Ryzen.
Neither CPU was particularly cheap: the 3.6 GHz (4 GHz boost) Ryzen 7 1800X retailed at $500, with the 0.2 GHz slower 1700X selling for $100 less. In part, AMD was keen to stop the perception of being the budget choice, but it was mostly because Intel was charging over $1,000 for their 8-core offering, the Core i7-6900K.
Zen took the best from all previous designs and melded them into a structure that focused on keeping the pipelines as busy as possible; and to do this, required significant improvements to the pipeline and cache systems. The new design dropped the sharing of L1/L2 caches, as used in Bulldozer, and each core was now fully independent, with more pipelines, better branch prediction, and greater cache bandwidth.
Reminiscent of the chip powering Microsoft and Sony's consoles, the Ryzen CPU was also a system-on-a-chip
Reminiscent of the chip powering Microsoft and Sony's consoles, the Ryzen CPU was also a system-on-a-chip; the only thing it lacked was a GPU (later budget Ryzen models included a GCN processor).
The die was sectioned into two so-called CPU Complexes (CCX), each of which were 4-core, 8-threads. Also packed into the die was a Southbridge processor -- the CPU offered controllers and links for PCI Express, SATA, and USB. This meant motherboards, in theory, could be made without an SB but nearly all did, just to expand the number of possible device connections.
All of this would be for nothing if Ryzen couldn't perform, and AMD had a lot to prove in this area after years of playing second fiddle to Intel. The 1800X and 1700X weren't perfect: as good than anything Intel had for professional applications, but slower in games.
AMD had other cards to play: a month after the first Ryzen CPUs hit the market, came 6 and 4-core Ryzen 5 models, followed two months later by 4-core Ryzen 3 chips. They performed against Intel's offerings in the same manner as their larger brothers, but they were significantly more cost effective.
And then came the aces -- the 16-core, 32-thread Ryzen Threadripper 1950X (with an asking price of $1,000) and the 32-core, 64-thread EPYC processor for servers. These behemoths comprised two and four Ryzen 7 1800X chips, respectively, in the same package, utilizing the new Infinity Fabric interconnect system to shift data between the chips.
In the space of six months, AMD showed that they were effectively targeting every x86 desktop market possible, with a single, one-size-fits-all design.
A year later, the architecture was updated to Zen+, which consisted of tweaks in the cache system and switching from GlobalFoundries' venerable 14LPP process -- a node that was under from Samsung -- to an updated, denser 12LP system. The CPU dies remained the same size, but the new fabrication method allowed the processors to run at higher clock speeds.
AMD launched Zen 2: this time the changes were more significant and the term chiplet became all the rage
Another 12 months after that, in the summer of 2019, AMD launched Zen 2. This time the changes were more significant and the term chiplet became all the rage. Rather than following a monolithic construction, where every part of the CPU is in the same piece of silicon (which Zen and Zen+ do), the engineers separated in the Core Complexes from the interconnect system.
The former were built by TSMC, using their N7 process, becoming full dies in their own right -- hence the name, Core Complex Die (CCD). The input/output structure was made by GlobalFoundries, with desktop Ryzen models using a 12LP chip, and Threadripper & EPYC sporting larger 14 nm versions.
The chiplet design will be retained and refined for Zen 3, currently penned for release late in 2020. We're not likely to see the CCDs break Zen 2's 8-core, 16-thread layout, instead it'll be a similar improvement as seen with Zen+ (i.e. cache, power efficiency, and clock speed improvements).
It's worth taking stock with what AMD achieved with Zen. In the space of 8 years, the architecture went from a blank sheet of paper to a comprehensive portfolio of products, containing $99 4-core, 8-thread budget offerings through to $4,000+ 64-core, 128-thread server CPUs.
AMD's finances have changed dramatically as well: from losses and debts running into the billions, AMD is now on track to clear its loans and post an operating income in excess of $600 million, within the next year. While Zen may not be the sole factor in the company's financial revival, it has helped enormously.
AMD's graphics division has seen similar changes in fortune -- in 2015, the section was given full independence, as the Radeon Technologies Group (RTG). The most significant development from their engineers came in the form of RDNA, a significant reworking of GCN. Changes to the cache structure, along with adjustments to the size and grouping of the compute units, shifted the focus of the architecture directly towards gaming.
The first models to use this new architecture, the Radeon RX 5700 series, demonstrated the design's serious potential. This was not lost on Microsoft and Sony, as they both selected Zen 2 and the updated RDNA 2, to power their forthcoming new Xbox and PlayStation 5 consoles.
AMD is quantifiably back to where it was in the Athlon 64 days in terms of architecture development and technological innovation. They rose to the top, fell from grace, and like a beast from mythology, created their own rebirth from the ashes.
Although the Radeon Group hasn't enjoyed the same level of success as the CPU division, and their graphics cards are perhaps still seen as the "value option," AMD is quantifiably back to where it was in the Athlon 64 days in terms of architecture development and technological innovation. They rose to the top, fell from grace, and like a beast from mythology, created their own rebirth from the ashes.
Looking ahead with caution
It's perfectly suitable to ask a simple question about AMD: could they return to the dark days of dismal products and no money?
Even if 2020 proves to be an excellent year for AMD and positive Q1 financial results show a 40% improvement to the previous year, $9.4 billion of revenue still puts them behind Nvidia ($10.7 billion in 2019) and light years away from Intel ($72 billion). The latter has a much larger product portfolio, of course, and its own foundries, but Nvidia's income is reliant almost entirely on graphics cards.
It's clear that both revenue and operating income need to grow, in order to fully stabilize AMD's future -- so how could this be achieved? The bulk of their income is from what they call the Computing and Graphics segment, i.e. Ryzen and Radeon sales. This will undoubtedly continue to improve, as Ryzen is very competitive and the RDNA 2 architecture will provide a common platform for games that work as well on PCs as they do on next-generation consoles.
Intel's latest desktop CPUs hold an ever decreasing lead in gaming. They also lack the breadth of features that Zen 3 will offer. Nvidia holds the GPU performance crown, but faces stiff competition in the mid-range sector from Radeons. It's perhaps nothing more than a coincidence, but even though RTG is a fully independent division of AMD, its revenue and operating income are grouped with the CPU sector -- this suggests that their graphics cards, while popular, do not sell in the same quantities as their Ryzen products do.
Possibly a more pressing issue for AMD is that their Enterprise, Embedded and Semi-Custom segment accounted for just under 20% of the Q1 2020 revenue, and ran at an operating loss. This may be explained by the fact that current-gen Xbox and PlayStation sales have stagnated, in light of the success of Nintendo's Switch and forthcoming new models from Microsoft and Sony. Intel has also utterly dominated the enterprise market and nobody running a multi-million dollar datacenter is going to throw it all out, just because an amazing new CPU is available.
But this could change over the next couple of years, partly through the new game consoles, but also from an unexpected alliance. Nvidia, of all companies, picked AMD over Intel as the choice of CPU for their new deep learning/AI compute clusters, the DGX 100. The reason is straightforward: the EPYC processor has more cores and memory channels, and faster PCI Express lanes than anything Intel has to offer.
If Nvidia is happy to use AMD's products, others will certainly follow suit. AMD will have to keep climbing a steep mountain, but today it'd appear they have the right tools for the job. As TSMC continues to tweak and refine its N7 process node, all AMD chips made using the process are going to be incrementally better, too.
Looking forward, there are a few areas within AMD that could use genuine improvement. One such area is marketing. The 'Intel Inside' catchphrase and jingle have been ubiquitous for over 30 years, and while AMD spends some money promoting Ryzen, ultimately they need makers such as Dell, HP, and Lenovo to sell units sporting their processors in the same light and specifications, as they do with Intel's.
On the software side, there's been plenty of work on applications that enhance users' experience, such as Ryzen Master, but it was just recently that Radeon drivers were having widespread problems. Gaming drivers can be hugely complex, but the quality of them can make or break the reputation of a piece of hardware.
AMD is currently in the strongest position that they've ever been in its 51-year history. With the ambitious Zen project showing no signs of hitting any limits soon, the company's phoenix-like rebirth has been a tremendous success. They're not at the top of the mountain though, and perhaps for the better. It's said that history always repeats itself, but let's hope that this doesn't come to pass. A healthy and competitive AMD, fully able to meet Intel and Nvidia head-on, only brings benefits to users.
What are your thoughts on AMD and its trials and tribulations -- did you own a K6 chip, or perhaps an Athlon? What's your favorite Radeon graphics card? Which Zen-based processor are you most impressed by? Share your comments in the section below.