Intel CEO laments Nvidia’s ‘extraordinarily lucky’ AI dominance, claims it coulda-woulda-shoulda have been Intel::Intel CEO Pat Gelsinger has taken a shot at his main rival in high performance computing, dismissing Nvidia’s success in providing GPUs for AI modelling as “extraordinarily lucky.” Gels

  • monkeyman512@lemmy.world
    link
    fedilink
    English
    arrow-up
    47
    ·
    7 months ago

    Having read the article it sounds like Pat is more complaining that Intel would have been positioned to milk the AI cow if the previous CEOs weren’t fucking idiots.

    • nomecks@lemmy.world
      link
      fedilink
      English
      arrow-up
      18
      ·
      7 months ago

      He’s right. Intel dumped their Tick Tock fast development cycle and let their product lines languish. Real stupid way to generate short term investor gains.

    • trolololol@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      ·
      7 months ago

      Ya sure he woulda coulda done different if he were in their places, because he is super hindsight man

    • CmdrShepard42@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      7 months ago

      Well Bob Swan did dump $20 billion on a single stock buyback that only elevated the share price for like 2 weeks and Brian Krzanitch saw delay after delay on 14nm which resulted in turning their two year lead into a two year defecit. There’s definitely truth to Pat’s words.

  • pdxfed@lemmy.world
    link
    fedilink
    English
    arrow-up
    30
    arrow-down
    2
    ·
    7 months ago

    I think it’s shorter to name the things Intel did strike gold on than the boats they missed despite all kinda of advantages.

    They missed the entire mobile market processor explosion, inevitable as it was since 2009 with Android launch(or 2007 if you want to say they should have seen what Apple was doing and thought they could compete, which if you’re an exec at Intel you should have).

    They missed cloud computing.

    Classic “big company hires/keeps overpaid check drawers instead of those with finger on the pulse”. Intel deserves the very little innovation, success and relevancy theyve had post 86.

    Bought my first AMD computer this year, an and 6800 Ryzen 7 with an on proc 680m gpu that is equivalent of ~ Nvidia 2050 discrete card. Game over for Intel.

    • Ottomateeverything@lemmy.world
      link
      fedilink
      English
      arrow-up
      16
      arrow-down
      6
      ·
      7 months ago

      Bought my first AMD computer this year, an and 6800 Ryzen 7 with an on proc 680m gpu that is equivalent of ~ Nvidia 2050 discrete card. Game over for Intel.

      While the rest of your post is logical, this is insane cope. No one is buying integrated graphics for gaming. 2050s are a joke in terms of power - you’re talking about a 2 year old budget mobile gpu… If anything this is basically a “I need to do some photoshop but don’t want a dedicated gpu on my laptop” type card. Intel has never given a fuck about mobile graphics. Their offerings have always been “serviceable, but get a real gpu if you want one”. Laptops are arguably better with ARM so there’s competition there…

      Intel is still selling their bread and butter and still has a huge stranglehold on their core market. Claiming “game over” because of an off case of an offshoot of one of their secondary markets is hugely overreacting.

      • rambaroo@lemmy.world
        link
        fedilink
        English
        arrow-up
        10
        arrow-down
        2
        ·
        7 months ago

        Plenty of people game on integrated graphics and a 2050 is damn powerful for an integrated card. Not everyone is a hyper nerd who builds their own PC or pays big bucks for a good gaming laptop. There are tons of casual players using integrated GPUs

  • Shadywack@lemmy.world
    link
    fedilink
    English
    arrow-up
    25
    arrow-down
    3
    ·
    7 months ago

    From the article

    Gelsinger explained how he thinks Nvidia and its CEO, Jensen Huang, just happened to be in the right place at the right time.

    Well, no shit. That’s how ALL rich people got rich, luck.

  • TheGrandNagus@lemmy.world
    link
    fedilink
    English
    arrow-up
    15
    ·
    edit-2
    7 months ago

    I’m becoming tired of Pat’s whining every other week. You don’t hear Lisa Su crying about Nvidia (or Intel for that matter) just being lucky, spouting this woe is me bullshit. You hear her, and Jensen, shutting up unless they have something meaningful to show.

    I’ve noticed this in a lot of Intel slides and presentations recently - they talk more about their competitor’s products than they do their own!

    Perhaps if intel had branched out into GPUs earlier than they did, while they still had mountains of money to do so, they could’ve leveraged the AI boom.

    But what did they do? They spaffed billions up the wall on stupid acquisitions like fucking McAfee Antivirus.

    Rather than swallowing their pride and using their manufacturing and CPU design skills to make ARM mobile CPUs, they engaged in the stupid decision to pay phone manufacturers to use x86 CPUs that simply weren’t efficient enough for mobile use. Needless to say, they all flopped hard.

    They let their foundries and design teams rot to the extent that AMD, a dying company, was able to surpass them.

    Now they’re struggling to release CPUs without them being 6 months to a year late. Sometimes they don’t release at all (where is desktop Meteor Lake, Intel?); that’s their bread and butter product, FFS! If Intel can’t do CPUs, what can they do?!

    Intel deserves everything they’re getting right now. Them getting left behind in AI is a problem entirely of their own making.

  • Eager Eagle@lemmy.world
    link
    fedilink
    English
    arrow-up
    14
    arrow-down
    4
    ·
    7 months ago

    If anything it should’ve been AMD. Intel is barely keeping up with the CPU competition these days.

    • hansl@lemmy.world
      link
      fedilink
      English
      arrow-up
      10
      arrow-down
      4
      ·
      7 months ago

      Not really. ATI were always “G is for graphics” and built video games cards. They never really saw the potential (nor did they have the resources anyway) for GPGPU, which is why NVIDIA had a huge first-player advantage (CUDA is 16 years old, 2 years before AMD acquired ATI). When AMD bought them it was already very late.

      Then AMD wanted to build cards for people to buy while NVIDIA was more than happy selling overpriced cards to crypto miners.

      OpenCL was an ambitious project that was too big and too open for what was capable from the Khronos group. Vulkan was too late.

      Intel could have done it but IIRC the CEO at that time (can’t remember the name) didn’t want to diversify their products after Itanium was a failure. They just doubled down on CPU.

      • Eager Eagle@lemmy.world
        link
        fedilink
        English
        arrow-up
        4
        ·
        edit-2
        7 months ago

        They never really saw the potential (nor did they have the resources anyway) for GPGPU

        Maybe ATI, which ended in 2010.

        AMD launched ROCm in 2016, after the first AI boom of 2012, but before GANs and transformers exploded. In recent years they’re better positioned in than Intel ever was.

      • bruhduh@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        7 months ago

        Brother, ati radeon hd 5000 series was basically vliw architecture single board computer in gpu package, and that was before amd bought radeon

      • TheGrandNagus@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        7 months ago

        Disagree. GCN cards were incredibly compute focused.

        Shit, AMD even invented HBM memory because they saw the value in ridiculously high bandwidth, dense, energy efficient memory in data centre applications. HBM is still used today in the enterprise market.

        AMD’s problem was that they had no money at the time and couldn’t build out their software ecosystem like Nvidia could - they had to bank on just getting the ball rolling and open sourcing their efforts in the hope that others would contribute, which didn’t happen to the extent that they’d have liked, especially when Nvidia with their mountains of cash could just pump out CUDA and flood universities with free GPUs to get them hooked in the Nvidia software stack.

    • EvergreenGuru@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      arrow-down
      1
      ·
      7 months ago

      Amd dropped the ball when it came to software and has now separated their GPU architecture so that they only have enterprise cards for data science. NVIDIA got in early and made CUDA default among all product lineups so that consumer cards could be used as entry-level cards by hobbyists. While it would’ve been nice to see more competition, the only company taking this space seriously has been nvidia.