Workers should learn AI skills and companies should use it because it’s a “cognitive amplifier,” claims Satya Nadella.

in other words please help us, use our AI

  • kescusay@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    15 days ago

    “Cognitive amplifier?” Bullshit. It demonstrably makes people who use it stupider and more prone to believing falsehoods.

    I’m watching people in my industry (software development) who’ve bought into this crap forget how to code in real-time while they’re producing the shittiest garbage I’ve laid eyes on as a developer. And students who are using it in school aren’t learning, because ChatGPT is doing all their work - badly - for them. The smart ones are avoiding it like the blight on humanity that it is.

    • wizardbeard@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      15 days ago

      As evidence: How the fuck is a company as big as Microsoft letting their CEO keep making such embarassing public statements? How the fuck has he not been forced into more public speaking training by the board?

      This is like the 4th “gaffe” of his since the start of the year!

      You don’t usually need “social permission” to do something good. Mentioning that is at best, publicly stating that you think you know what’s best for society (and they don’t). I think the more direct interpretation is that you’re openly admitting you’re doing the type of thing that you should have asked permission for, but didn’t.

      This is past the point of open desperation.

    • devfuuu@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      15 days ago

      And they are all getting dependent and addicted to something that is currently almost “free” but the monetization of it all will soon come in force. Good luck having the money to keep paying for it or the capacity to handle all the advertisement it will soon start to push out. I guess the main strategy is manipulate people into getting experience with it with these 2 or 3 years basically being equivalent to a free trial and ensuring people will demand access to the tools from their employees which will pay from their pockets. When barely anyone is able to get their employers to pay for things like IDEs… Oh well.

    • hushable@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      15 days ago

      I’m watching people in my industry (software development) who’ve bought into this crap forget how to code in real-time while they’re producing the shittiest garbage I’ve laid eyes on as a developer.

      I just spent two days fixing multiple bugs introduced by some AI made changes, the person who submitted them, a senior developer, had no idea what the code was doing, he just prompted some words into Claude and submitted it without checking if it even worked, then it was “reviewed” and blindly approved by another coworker who, in his words, “if the AI made it, then it should be alright”

    • floofloof@lemmy.ca
      link
      fedilink
      English
      arrow-up
      1
      ·
      15 days ago

      I’ve been programming professionally for 25 years. Lately we’re all getting these messages from management that don’t give requirements but instead give us a heap of AI-generated code and say “just put this in.” We can see where this is going: management are convincing themselves that our jobs can be reduced to copy-pasting code generated by a machine, and the next step will be to eliminate programmers and just have these clueless managers. I think AI is robbing management of skills as well as developers. They can no longer express what they want (not that they were ever great at it): we now have to reverse-engineer the requirements from their crappy AI code.

      • Null User Object@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        15 days ago

        but instead give us a heap of AI-generated code and say “just put this in.”

        we now have to reverse-engineer the requirements from their crappy AI code.

        It may be time for some malicious compliance.

        Don’t reverse engineer anything. Do as your told and “just put this in” and deploy it. Everything will break and management will explode, but now you’ve demonstrated that they can’t just replace you with AI.

        Now explain what you’ve been doing (reverse engineering to figure out their requirements), but that you’re not going to do that anymore. They need to either give you proper requirements so that you can write properly working code, or they give you AI slop and you’re just going to “put it in” without a second thought.

        You’ll need your whole team on board for this to work, but what are they going to do, fire the whole team and replace them with AI? You’ll have already demonstrated that that’s not an option.

  • Fizz@lemmy.nz
    link
    fedilink
    English
    arrow-up
    1
    ·
    15 days ago

    Its funny because this is an admission that hes not actually done anything interesting which is a complete pivot from the last two years of him screaming about how good AI is for the last 2 years.

  • fyrilsol@kbin.melroy.org
    link
    fedilink
    arrow-up
    1
    ·
    15 days ago

    “social permission”?

    Society didn’t even permit you and others to spread AI onto everyone to begin with.

  • FlashMobOfOne@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    15 days ago

    “Social permission” is one term for it.

    Most people don’t realize this is happening until it hits their electric bills. Microslop isn’t permitted to steal from us. They’re just literal thieves and it takes time for the law to catch up.

    • 100@fedia.io
      link
      fedilink
      arrow-up
      1
      ·
      15 days ago

      you will enjoy your chatbot that confidently tells lies while electricity bill goes up by 50% and the nearby datacentres try to make the next model not use em-dashes

      • floofloof@lemmy.ca
        link
        fedilink
        English
        arrow-up
        1
        ·
        15 days ago

        As a long-time user of the em-dash I’m pissed off that my usual writing style now makes people think I used AI. I have to second-guess my own punctuation and paraphrase.

  • SeeMarkFly@lemmy.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    15 days ago

    So…he has something USELESS and he wants everybody to FIND a use for it before HE goes broke?

    I’ll get right on it.

    • glimse@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      15 days ago

      I was expecting something much worse but to me it deels like he’s saying “we, the people working on this stuff, need to find real use cases that actually justifies the expense” which is…pretty reasonable

      Not defending him or Microsoft at all here but it sounds like normal business shit, not a CEO begging users to like their product

      • TrousersMcPants@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        15 days ago

        I mean, it would be a lot more reasonable if the entire tech industry hadn’t gone absolutely 100% all-in on investing billions and billions of dollars into the technology before realizing that they didn’t have any use cases to justify that investment.

        • glimse@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          15 days ago

          Oh you don’t have to convince me it was a mistake, his comment just wasn’t what it’s being made out to be. “Find a use for it or we’re fucked” is a lot different than “please use our product or we’re fucked”

            • glimse@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              15 days ago

              I dunno, there’s a pretty big distinction. “Please use our shitty thing” vs “Please make our shitty thing better so people want to use it”

              It’s placing “blame” on the industry leaders for failing to make something useful

              • msage@programming.dev
                link
                fedilink
                English
                arrow-up
                1
                ·
                14 days ago

                But HE is the industry leader.

                It’s not like it wasn’t him that sunk every successful product they still had in the last couple of years.

  • Aceticon@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    14 days ago

    AI isn’t at all reliable.

    Worse, it has a uniform distribution of failures in the domain of seriousness of consequences - i.e. it’s just as likely to make small mistakes with miniscule consequences as major mistakes with deadly consequences - which is worse than even the most junior of professionals.

    (This is why, for example, an LLM can advise a person with suicidal ideas to kill themselves)

    Then on top of this, it will simply not learn: if it makes a major deadly mistake today and you try to correct it, it’s just as likely to make a major deadly mistake tomorrow as it would be if you didn’t try to correct it. Even if you have access to actually adjust the model itself, correcting one kind of mistake just moves the problem around and is akin to trying to stop the tide on a beach with a sand wall - the only way to succeed is to have a sand wall for the whole beach, by which point it’s in practice not a beach anymore.

    You can compensate for this by having human oversight on the AI, but at that point you’re just back to having to pay humans for the work being done, so now instead of having to the cost of a human to do the work, you have the cost of the AI to do the work + the cost of the human to check the work of the AI and the human has to check the entirety of the work just to make sure since problems can pop-up anywere, take and form and, worse, unlike a human the AI work is not consistent so errors are unpredictable, plus the AI will never improve and it will never include the kinds of improvements that humans doing the same work will over time discover in order to make later work or other elements of the work be easier to do (i.e. how increase experience means you learn to do little things to make your work and even the work of others easier).

    This seriously limits the use of AI to things were the consequences of failure can never be very bad (and if you also include businesses, “not very bad” includes things like “not significantly damage client relations” which is much broader than merely “not be life threathening”, which is why, for example, Lawyers using AI to produce legal documents are getting into trouble as the AI quotes made up precedents), so mostly entertainment and situations were the AI alerts humans for a potential situation found within a massive dataset and if the AI fails to spot it, it’s alright and if the AI incorrectly spots something that isn’t there the subsequent human validation can dismiss it as a false positive (so for example, face recognition in video streams for the purpose of general surveillance, were humans watching those video streams are just or more likely to miss it and an AI alert just results in a human checking it, or scientific research were one tries to find unknown relations in massive datasets)

    So AI is a nice new technological tool in a big toolbox, not a technological and business revolution justifying the stock market valuations around it, investment money sunk into it or the huge amount of resources (such as electricity) used by it.

    Specifically for Microsoft, there doesn’t really seem to be any area were MS’ core business value for customers gains from adding AI, in which case this “AI everywhere” strategy in Microsoft is an incredibly shit business choice that just burns money and damages brand value.

        • veni_vedi_veni@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 days ago

          Yea that’s my obscure reference, it was a whole scandal because nobody except him thought it was a good idea to represent them on Fox, who specifically reached out to him (her? the guy was something non-cis) because they knew he was a bozo

  • NutWrench@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    15 days ago

    The whole point of “AI” is to take humans OUT of the equation, so the rich don’t have to employ us and pay us. Why would we want to be a part of THAT?

    AI data centers are also sucking up all the high quality GDDR5 ram on the market, making everything that relies on that ram ridiculously expensive. I can’t wait for this fad to be over.

    • danielton1@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      15 days ago

      Not to mention the water depletion and electricity costs that the people who live near AI data centers have to deal with, because tech companies can’t be expected to be responsible for their own usage.

    • FilthyShrooms@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      15 days ago

      Denial: “AI will be huge and change everything!”

      Anger: “noooo stop calling it slop its gonna be great!”

      Bargaining: “please use AI, we spent do much money on it!”

      Depression: companies losing money and dying (hopefully)

      Acceptance: everyone gives up on it (hopefully)

      • rumba@lemmy.zip
        link
        fedilink
        English
        arrow-up
        1
        ·
        15 days ago

        Acceptance: It will be reduced to what it does well and priced high enough so it doesn’t compete with equivalent human output. Tons of useless hardware will flood the market, china will buy it back and make cheap video cards from the used memory.