• r00ty@kbin.life
    link
    fedilink
    arrow-up
    6
    ·
    1 year ago

    Yes, and businesses thinking they can drop their developers for chatgpt like tech in the future should (they won’t, but they should) consider this. AI goes to pot very quickly without human input.

    • lemmyvore@feddit.nl
      link
      fedilink
      English
      arrow-up
      5
      ·
      edit-2
      1 year ago

      It’s not even on the map. Most of the businesses who think they can replace anybody with LLMs are thinking about subscribing to a LLM that’s been trained and maintained by someone else. Which of course involved giving that someone the upper hand and letting them dictate terms.

      Anybody who tried making their own model knows it’s tough, grueling work.

      So these businesses take the easy way out and will give that someone their data (and break privacy and regulations in the process) and use the data that comes out of the LLM with no regard whatsoever about where and how it’s been sourced and what legal implications that might have for themselves.

      If you add the fact the LLM owner usually makes you sign a contract that gives basically no guarantees, you have the recipe for a very fine mess.

      I still can’t wrap my head around for example how can any software company let or even goad its programmers to use Copilot in good faith, with no idea where the code is coming from and what’s the copyright status. Leaving aside the fact Microsoft is currently being sued for this exact problem.

    • EatMyDick@lemmy.world
      link
      fedilink
      arrow-up
      1
      arrow-down
      6
      ·
      1 year ago

      Guys stop this non-sense. That’s not how it works.

      They hire less new developers. There will be less people doing the work. Idiots who don’t learn to use the tech will be left behind. This is already happening.