Become a fan of Slashdot on Facebook

 



Forgot your password?
typodupeerror
×
Intel Linux Hardware

Intel's Failed 64-bit Itanium CPUs Die Another Death as Linux Support Ends (arstechnica.com) 78

Officially, Intel's Itanium chips and their IA-64 architecture died back in 2021, when the company shipped its last processors. But failed technology often dies a million little deaths. From a report: To name just a few: Itanium also died in 2013, when Intel effectively decided to stop improving it; in 2017, when the last new Itanium CPUs shipped; in 2020, when the last Itanium-compatible version of Windows Server stopped getting updates; and in 2003, when AMD introduced a 64-bit processor lineup that didn't break compatibility with existing 32-bit x86 operating systems and applications.

Itanium is dying another death in the next version of the Linux kernel. According to Phoronix, all code related to Itanium support is being removed from the kernel in the upcoming 6.7 release after several months of deliberation. Linus Torvalds removed some 65,219 lines of Itanium-supporting code in a commit earlier this week, giving the architecture a "well-earned retirement as planned."

This discussion has been archived. No new comments can be posted.

Intel's Failed 64-bit Itanium CPUs Die Another Death as Linux Support Ends

Comments Filter:
  • Itanic (Score:5, Funny)

    by eneville ( 745111 ) on Friday November 03, 2023 @01:08PM (#63977404) Homepage
  • But Intel vastly underestimated the difficulties involved in making it work as imagined
    I admire their bravery and creativity and remind critics that many innovative new ideas fail
    We need more companies willing to take expensive chances on promising, but difficult, ideas

    • by Z80a ( 971949 )

      I think the most recent example of it is the apple M1, where they did bet on a very wide execution unit.
      It's said in some places that every extra parallel instruction you add to a super scalar chip, the size of the decode logic must double if you want to keep the same performance of the decode logic.
      And they went from the standard now standard 4 parallel instructions to 8 parallel instructions, which would require either a 16 times bigger decode logic, or a decode logic that is worse per instruction.
      My bet

      • Apple also control the software ecosystem so they could provide a seamless, and in many cases, a transparent upgrade path. Most old software runs, and universal binaries hid some complexity. I also donâ(TM)t remember Intel working so hard with third parties had software ported and optimised at launch time. Likewise, AMD provided a compatible solution so users had no problem with the transition.

        • HP-UX on Itanium ran PA-RISK processes *very* well. As seamless as MacOS as far as I can tell. Unfortunately HP-UX did not have much market share and Microsoft never seemed to really buy into Itanium. Intel was very dependent on Microsoft.
        • by tlhIngan ( 30335 )

          Apple also control the software ecosystem so they could provide a seamless, and in many cases, a transparent upgrade path. Most old software runs, and universal binaries hid some complexity. I also donÃ(TM)t remember Intel working so hard with third parties had software ported and optimised at launch time. Likewise, AMD provided a compatible solution so users had no problem with the transition.

          Apple had an easier time because even the M3 today is still ARM compatible, so every ARM AArch64 binary produc

      • >It's said in some places that every extra parallel instruction you add to a super scalar chip, the size of the decode logic must double if you want to keep the same performance of the decode logic.

        As far as I know, this simply isn't true. Also, the decode logic is small compared to the rest of the CPU, so even if it is true, it wouldn't be such a big deal.
        I'm not up on implementation of modern ARMs, so I might be wrong, but I work for a CPU vendor and in my domain I think it's right for current generati

      • by edwdig ( 47888 )

        Apple's M CPUs include a GPU and all the system RAM on the processor. The decode logic has got to be such a ridiculously small percentage of the chip that it's not even worth thinking about the size of it.

        • In any modern amd64 processor instruction decode is so small that even if you doubled the size you'd still have to squint to see it. And that has to handle x86 and amd64!

    • by crgrace ( 220738 )

      I'm not sure it was a good idea at the time. When it was being developed, there were a lot of discussions about Itanium here and elsewhere and the general consensus was that going to a VLIW processor was premature because the compilers were not yet up to the task and there wasn't a real path forward that would bring a significant performance improvement.

      It was sad watching HP retired PA-RISC to go all in on Itanium. The PA-RISC chips had huge (for the time) caches and were the best architecture (in my opini

      • HP did a fabulous job making PA-RISC software run unmodified on Itanium. I don't know about performance. But the experience was quite good. Unfortunately, HP-UX itself was never able to build up enough market share to support a processor family. (Which is likely why HP was willing to retire PA-RISC)
      • by rahmrh ( 939610 ) on Friday November 03, 2023 @03:28PM (#63977836)

        It was a good idea at the time, but Intel did not know AMD was busy and was going to release a 64-bit cpu with 18months of the first decent Itaniums being shipped.

        AMD's 64-bit cpu sneak attack had a lot to do with Itanium not being viable. It removed most of the Itanium advantages and left Itanium with the limited production price disadvantage that it was never able to overcome. I am not going to buy an Itanium that costs several thousand more and always uses full power over an 64-bit AMD that uses way less power at idle.

        The 2004 Itaniums had ZERO power savings. I put a power meter on multiple boxes and they used the same wattage +-10 watts no matter if the OS was idle or if it was running a benchmark. So there is the extra power costs and cooling costs. And compared to an equivalent dual socket AMD the itaniums also used more power running a benchmark vs the AMD with the same benchmark.

        • It was a good idea at the time, but Intel did not know AMD was busy and was going to release a 64-bit cpu with 18months of the first decent Itaniums being shipped.

          Given every RISC processor had 64-bit versions at the time, it's hard to imagine Intel didn't know 64-bit was inevitable. And I doubt Andy "only the paranoid survive" Grove didn't know AMD was feverishly working on 64-bit extensions.

          The 2004 Itaniums had ZERO power savings. I put a power meter on multiple boxes and they used the same wattage +-10 watts no matter if the OS was idle or if it was running a benchmark.

          I don't think that was a design goal. No one cared about server power draw until around 2006 or so. That's when power suddenly became important and the race to ever higher clock rates screeched to a halt.

          The goal if Itanium was always about cycle time and masking memory latency.

          • And I doubt Andy "only the paranoid survive" Grove didn't know AMD was feverishly working on 64-bit extensions.

            Interesting question. I believe they thought they could successfully partition the "enterprise" and desktop markets and continue to summon ridiculous margins from the business unit while containing fast desktop machines to gaming and other cheapifying software forces. The AMD64 ISA offered the exact opposite vision and we all see how that turned out. I thought it especially entertaining to watch Intel have to negotiate AMD64 licensing.

            I don't think that was a design goal.

            I noticed significant power savings starting with the rx2800 series serve

            • The main thing I remember about the Itanic was that any lack of memory alignment killed performance to a much greater degree than it did on other platforms. Alignment faults caused many wasted cycles and were/are easy to cause. I have to admit to liking this ISA the least of any I've had to work on. Their promises even started off sounding like BS, because it all depended on having the magic compiler complete the basic design goals... sooooo where was it? We never got it!

              Well. Having worked at HP at the time and listened into many break room gripe sessions about Intel, the common complaints were about the Intel Value Engine (IVE). Intel demanded Itanium include it, HP strongly preferred transpiling and software emulation. The IVE let an Itanium natively run x86 code. The interrupt mechanism in particular caused hissy fits for the people designing the actual cores.

              While mis-aligned memory loads could have been a problem (maybe that's what was causing all the interrupts), I'm

              • Having worked at HP at the time and listened into many break room gripe sessions

                How long were you there? I'm ex-IBM, ex-SGI, but never HP. Did you catch the the whole internal drama around the x86 port of HP-UX? I'd love to hear that story. Have any opinions on where HP-UX goes from here, if anywhere?

                I'm pretty confident the hardware/BIOS guys could have figured that out.

                I bet you are right. However, I think the politics and business-side of things got so complicated and legally fraught with peril that the technical issues were secondary, I think.

                • How long were you there? I'm ex-IBM, ex-SGI, but never HP. Did you catch the the whole internal drama around the x86 port of HP-UX? I'd love to hear that story. Have any opinions on where HP-UX goes from here, if anywhere?

                  I left HP-UX development around 1998. 11.0 was in flight at the time. I got out long before we started working on x86 (not really "we", I was at another company by then). When I left, getting HP-UX running on native Itanium was the focus.

                  What's the future of HP-UX? Not much would be my guess. At most, open source the code and let people hack away at it. Linux has long since eclipsed HP-UX in performance, platforms, features, and application support. There's no reason to have a gratuitously different niche c

                  • Ah, so you were there when the PA-RISC was just starting to wane and the big push to the Itanium, gotcha. I think the x86-port foolery happened closer to 2001-2003, IIRC. I'd like to see HP-UX open-sourced (not as much as I'd like IRIX open source, but still). If OpenVMS is any indication, they probably would rather sell it off to some small company for maintainership, like VSI. However, there is so much 3rd party IP in there, it'd be a tough slog.
          • by edwdig ( 47888 )

            Given every RISC processor had 64-bit versions at the time, it's hard to imagine Intel didn't know 64-bit was inevitable. And I doubt Andy "only the paranoid survive" Grove didn't know AMD was feverishly working on 64-bit extensions.

            The general vibe at the time was that x86 was awful and that extending it further wasn't viable. Pretty much everyone assumed something would replace it, and that was the big battleground. Intel might have known that AMD was looking into it, but I don't think they took the threat seriously.

            The goal if Itanium was always about cycle time and masking memory latency. Tons of architectural decisions were made with the assumption memory was slow and so you needed ways to hide it. I don't think we (I worked on HP-UX at the time) had any idea that multi-core would take off like it did. We implicitly assumed a single thread would be running and that the ISA/control unit would need to keep enough balls in the air. Turned out throwing transistors at many cores worked much better.

            It's fascinating to look back and see how wrong people were about that. This is the era of the Cell Processor, which only really got used in the PlayStation 3. It's such a weird design - a traditional PowerPC core surroun

            • It's fascinating to look back and see how wrong people were about that. This is the era of the Cell Processor, which only really got used in the PlayStation 3. It's such a weird design - a traditional PowerPC core surrounded by a lot of very fast but very limited mini cores. People were starting to realize that multiple cores was the future, but they hadn't yet realized that the cores would be full cores. By the time Cell made it to production, people had already mostly realized it was a bad idea.

              It's much

              • by edwdig ( 47888 )

                I think the original idea was that the cells would act as the GPU on the PS3, but once they started testing it, they realized it couldn't come close to the performance needed. They had rework the design to include a traditional GPU, and that's a large part of why the PS3 was so much more expensive than the Xbox 360. (BluRay being the other reason)

                It hit a weird spot somewhere in between modern CPUs and GPUs. In the end it wasn't good enough at either role. But if you had just the right type of workload, it

            • The general vibe at the time was that x86 was awful and that extending it further wasn't viable. Pretty much everyone assumed something would replace it,

              Everyone but Intel and AMD.

              It's fascinating to look back and see how wrong people were about that. This is the era of the Cell Processor, which only really got used in the PlayStation 3. It's such a weird design - a traditional PowerPC core surrounded by a lot of very fast but very limited mini cores.

              And yet, now we're seeing mass market CPUs with performance cores, low power cores, AI cores, graphics cores, all sorts of specialized hardware you turn on or off based on the load.

              I'm trying to remember the RISC V architecture. Does it allow for multiple special purpose cores in the architecture? Or is it silent on that, leaving it to system designer?

    • Re: (Score:3, Interesting)

      by Narcocide ( 102829 )

      Arguably what AMD succeeded at after them was significantly more difficult, but they still made it work and they were even able to make it do so cheaply. Intel rested on their laurels, confident they could just keep fudging benchmarks and bribing vendors to maintain market dominance. Keep in mind, in their plan, home users would never get 64-bit technology. They planned on ceding ground there so they could squeeze datacenter customers harder. This is unacceptable on a number of levels and I'm glad it ended

      • Intel did the same thing when moving from the 16 bit 286 to the 32 bit 386.

      • Arguably what AMD succeeded at after them was significantly more difficult, but they still made it work and they were even able to make it do so cheaply. Intel rested on their laurels, confident they could just keep fudging benchmarks and bribing vendors to maintain market dominance. Keep in mind, in their plan, home users would never get 64-bit technology. They planned on ceding ground there so they could squeeze datacenter customers harder. This is unacceptable on a number of levels and I'm glad it ended with a faceplant.

        I'm not sure it was more difficult, just a better idea.

        x86 was always known to be a complicated architecture, but it was king so people figured it out. Intel figured they'd make an awesome new architecture from scratch for Itanium, and maybe it was awesome, but the problem is that everyone else had to learn it from scratch.

        AMD just extended x86 and everybody went with that since it was easier.

        • I'm not sure it was more difficult, just a better idea.

          Well, it was definitely a better idea, but keep in mind that up until the day after AMD pulled this off, Intel was still telling everyone that making the jump to 64-bit while maintaining binary compatibility with 32-bit architectures was [literally impossible] and they were using this to justify a lot of their egregious pricing.

    • by garyisabusyguy ( 732330 ) on Friday November 03, 2023 @01:29PM (#63977472)

      What is sad is that there were far more capable architecture available at the time, like DEC ALPHA, and Intel pressured their customers to "capture and kill" ALPHA thinking that would give them a clear path to Itanium supremacy...

      • That's what I remember too. Itanic killed Alpha (through marketing, not performance), and now it's crashed into its iceberg. RIP Alpha. A pox on you Itanic!

      • by sjames ( 1099 )

        That's the part that's unforgivable. They steamrolled everything else flat hoping that if they were the only 64 bit game in town, they could sell it even if it wasn't as good.

        Thankfully AMD came in and ate their lunch.

        Intel didn't even make sure their own compiler could produce decent performing code on Itanic.

      • Alpha died because it wouldn't scale further without substantial rework.

        It was cheaper to put the good parts of the design into a PC processor, which AMD did.

        • Alpha had a planned roadmap of three decades and EV8 (developed but un-taped) would have included vector processing units, 2 cores and 4.0Ghz for 2003 production

          In addition, they used the HyperTransport interconnect that that continues to be used in multiple processors today from AMD, Apple and IBM [wikipedia.org]

          Interestingly enough, China has "reverse engineered" DEC ALPHA for Sunway supercomputer chip [nextplatform.com]

          You say funny things, but I do not think that you are trying to be funny

          • Alpha had a planned roadmap of three decades and EV8 (developed but un-taped) would have included vector processing units, 2 cores and 4.0Ghz for 2003 production

            I can draw you a map to a place that doesn't exist, too. They were having trouble even getting Alpha into the GHz range.

          • by edwdig ( 47888 )

            Alpha had a planned roadmap of three decades and EV8 (developed but un-taped) would have included vector processing units, 2 cores and 4.0Ghz for 2003 production

            I wouldn't put much weight into that. Intel's roadmap then was "We'll have a 10 GHz Pentium 4 in a couple years."

            That point in time was right before everyone realized that heat was going to be a way bigger limiting factor in chip design than it had been in the past.

      • PA-RISC was also killed off. I'm fuzzy on the history of Itanium, but apparently HP decided of off their own CPU design and get in bed with Intel, a move that confuses me greatly. PA-RISC wasn't spectacular, but it was competitive, and apparently had some very good SIMD support at the time.

    • Technically, it was a good idea. However comercially, Intel was a victim of it's own monopoly in the PC market, and by not being sufficiently backwards compatible. Of course, it also encountered development delays and eventually ended up being slower than a competing RISC chip with similar transistor counts. A snag Intel has had a long time - it likes the golden goose revenue generation of x86 family, but it also needs to innovate with something new, but in ways that don't harm the golden goose.

    • by jonadab ( 583620 )
      > But Intel vastly underestimated the difficulties involved in making it work as imagined

      The main thing they underestimated, was the extent to which customers value compatibility with existing firmly-established stuff. The entire computing world heard that Intel was working on a new 64-bit architecture and collectively went "Ok, so what's it compatible with?" And the answer was nothing, and that was pretty much the end of that.
  • All their plans are greedy, but this one was also particularly stupid and eventually the universe itself rejected it.

    The End.

  • ah, you know the rest....

  • Here's the actual, more complete merge comment:

    The ia64 architecture gets its well-earned retirement as planned,
    now that there is one last (mostly) working release that will be
    maintained as an LTS kernel.

    So it isn't like kernel support is going completely away. It just won't be included in future releases.

  • Outside of cluster use, who was running linux on Itanium boxes in production? Chances are you bought Itanium because you needed to run OpenVMS.

    • Or HPUX. I used to manage a farm of Superdomes loaded with Itanic CPUs.

      There was a time they kicked Sunâ(TM)s ass. The I/O on Superdome was far superior to the Sun Fire x800 line at the time.

      HPUX never caught on quite as hot as it should have. It had its quirks but man, once you had it humming, it just cruised.

      Part of me (a very small part) misses it.

      • In 2003 I actually helped put together an application for a state agency on an HP Superdome. Yep.. it was fast. Wicked fast I/O.
      • by rahmrh ( 939610 )

        We had a bad nickname for our Superdomes. We had 100's of HPUX machines 90% of which we never touched/repaired. But we had a few Stuperdomes have significant hardware issues more than a few times (random hardware died). I am pretty sure one of the oncalls touched each of ours at least 2x to either do a engineering change (based on a prior hardware defect/crash) or get it back working again because some piece of hardware died and would not let the machine come back up.

        Now, comparing the to the IBM hardwar

        • I think i failed to mention we implemented the PA/RISC Superdomes.. not the Itanium. We had the IBM Power Series too. An IBM Power HA Cluster was for the DMV, One HP SuperDome Pair was for DEM, A second HP SuperDome Pair was for DCYF. ahh.. he good old days. We never had an outage.
          • by rahmrh ( 939610 )

            I don't believe I ever used one of the PA/RISC superdomes. The Itanium ones were new design/engineering and troublesome. I do know that the PA/Risc workstations and such I managed 5-8 years earlier rarely gave us much trouble. We found out one of them had been our DNS server for the entire site and had never crashed in 5+ years. The DNS team figured out the risk from the single piece of hw/os and went to a "highly" reliable dual hardware/software setup and has proceeded to have multiple dns outages in

      • by HiThere ( 15173 )

        OK. And a very small part of me misses programming in Regal on the Harris computer. But it's a *REALLY* small part.

      • I had a contract to do ipsec between windows and hpsux on itanic. I found that the examples in the ipsec documentation were backwards regarding which end was which. Tried them verbatim and nothing worked. Reversed them and they worked fine. Hilarious.

  • It was dead way before 2013, HPE would just not admit it. By 2006 (outside of HPUX/HPE) no one bought them. The integer performance was not as good as an X86/AMD and they only really beat X86/AMD on certain floating point operations. The same speed of most applications on X86/AMD (motherboard + 2 cpus) was about 1/2 the price (about $1000+ less). And the Itaniums needed different power supplies and motherboard mounts so more $$ there too. The Itaniums cpus cost more and the motherboards cost more and

    • even in 2008 Itanium2 held database benchmarks for performance... for those that didn't care about price/performance but absolute numbers. I worked for place that sold the systems. After that Itanium2 went into decline.

  • by Required Snark ( 1702878 ) on Friday November 03, 2023 @05:48PM (#63978168)
    There was an unacknowledged monopolistic goal that set the agenda. At that time the Window/Intel de facto monopoly was in full swing, and the only real competition was AMD and a few minor players like Cyrix. Intel wanted complete world domination.

    By pushing a wildly incomparable architecture where they owned the IP, they intended to control the desktop until the end of time. All it would take is buy-in by Microsoft and with the depreciation of x86 for Windows and they would have succeeded. No other IC company would have the resources to complete in that situation.

    Unfortunately for Intel, and fortunately for everyone else on the planet, AMD did a good job of compatibly extending the x86 instruction set. The flop of the Itanium and success of AMD kept us from a stagnant inefficient computing culture that would have strangled innovation and kept computing expensive. In that repressive scenario it's not clear that an alternative like ARM would have had a chance, much less the current emergence of RISC-V.

    The same situation is happening right now with Google search. Their 90% market share and excessive profits are hindering web innovation. Online consumers are paying a toll to Google and receiving biased results driven by paid advertising. SEO is a polite way of saying extortion/bribery.

    At this time there is no other credible competition, despite the hype about AI driven search. Google can keep buying search dominance and still afford to make some AI enabled solution, and even if it isn't the best they can stay on top.

    Without intervention by US or EU regulators the situation will continue to deteriorate. No matter how much they claim to be doing a great job, no single company can realistically create an efficient, fair, and flexible future. Monopoly always ends up in failure, and that is the way we are headed.

    • Unfortunately for Intel, and fortunately for everyone else on the planet, AMD did a good job of compatibly extending the x86 instruction set.

      ...
      The same situation is happening right now with Google search.

      ...
      Without intervention by US or EU regulators the situation will continue to deteriorate.

      If this is the same situation, and the previous situation (with Intel) didn't need intervention, why would the situation now (with Google) require intervention?

      Arguably, the situation now is simpler, as it's much easier to launch The Next Big Thing as a software startup than the to get a new IC contender off the ground.

  • I was an Intel reseller back in the 1990s/early aughts. They'd have these get-togethers for us where they'd announce new products, lay out roadmaps, etc. And so much of what they touted never came to pass.

    I remember being told for a couple of years how RAMbus was going to provide significant speed increases over SDRAM; that it would initially be "a little" more expensive but would rapidly reach cost parity and ultimately be cheaper. The Pentium 4 "was designed for RAMbus" and later P4 chipsets wouldn't e

news: gotcha

Working...