OpenAI’s Mira Murati: “some creative jobs maybe will go away, but maybe they shouldn’t have been there in the first place” And you stole everything from creative people who provided free texts, images, forum answers, etc. To date, your company has refused to acknowledge any credit. Rich people truly live in their bubble and have zero sympathy for fellow human or their livelihood.

  • MudMan@fedia.io
    link
    fedilink
    arrow-up
    2
    arrow-down
    1
    ·
    4 months ago

    Every industrial transition generates that, though. Forget the Industrial Revolution, these people love to be compared to that. Think of the first transition to data-driven businesses or the gig economy. Yeah, there’s a chunk of people caught in the middle that struggle to shift to the new model in time. That’s why you need strong safety nets to help people transition to new industries or at least to give them a dignified retirement out of the workforce. That’s neither here nor there, if it’s not AI it’ll be the next thing.

    About the linear increase path, that reasoning is the same old Moore’s law trap. Every line going up keeps going up if you keep drawing it with the same slope forever. In nature and economics lines going up tend to flatten again at some point. The uncertainty is whether this line flattens out at “passable chatbots you can’t really trust” or it goes to the next step after that. Given what is out there about the pace of improvement and so on, I’d say we’re probably close to progress becoming incremental, but I don’t think anybody knows for sure yet.

    And to be perfectly clear, this is not the same as saying that all tech disruption is good. Honestly, I don’t think tech disruption has any morality of any kind. Tech is tech. It defines a framework for enterprise, labor and economics. Every framework needs regulation and support to make it work acceptably because every framework has inequalities and misbehaviors. You can’t regulate data capitalism the way you did commodities capitalism and that needed a different framework than agrarian societies and so on. Genies don’t get put back in bottles, you just learn to regulate and manage the world they leave behind when they come out. And, if you catch it soon enough, maybe you get to it in time to ask for one wish that isn’t just some rich guy’s wet dream.

    • nickwitha_k (he/him)@lemmy.sdf.org
      link
      fedilink
      arrow-up
      4
      ·
      4 months ago

      Think of the first transition to data-driven businesses or the gig economy.

      Just a clarification: the “gig economy” was not “new” in any way, just using new technology to skirt around labor laws and find loopholes in regulations in order to claw back profits that had been “lost” to things like pensions and health coverage.

      • MudMan@fedia.io
        link
        fedilink
        arrow-up
        2
        ·
        4 months ago

        Well, yeah, that’s what I’m talking about here, specifically. There was an application of technology that bypassed regulations put in place to manage a previous iteration of that technology and there was a period of lawlessness that then needed new regulation. The solutions were different in different places. Some banned the practice, some equated it with employees, some with contractors, some made custom legislation.

        But ultimately the new framework needed regulation just like the old framework did. The fiction that the old version was inherently more protected is an illusion created by the fact that we were born after common sense guardrails were built for that version of things.

        AI is the same. It changes some things, we’re gonna need new tools to deal with the things it changes. Not because it’s worse, but because it’s the same thing in a new wrapper.

    • octopus_ink@lemmy.ml
      link
      fedilink
      English
      arrow-up
      2
      ·
      4 months ago

      That’s why you need strong safety nets to help people transition to new industries or at least to give them a dignified retirement out of the workforce. That’s neither here nor there, if it’s not AI it’ll be the next thing.

      I agree with most of what you wrote in this paragraph, but we have no such strong safety nets. I don’t think the fact that it has happened previously is justification for creating those circumstances again now (or in the future) without concern for how it impacts people. We’re supposed to be getting better as time goes by. (not that we are by many other metrics I can see on a daily basis, but as you say that’s another conversation)

      Genies don’t get put back in bottles, you just learn to regulate and manage the world they leave behind when they come out. And, if you catch it soon enough, maybe you get to it in time to ask for one wish that isn’t just some rich guy’s wet dream.

      I also agree with this.

      But, I find there is plenty of justification to push back and try to slow the proliferation of AI in certain areas while our laws and morality try to catch up.

      • MudMan@fedia.io
        link
        fedilink
        arrow-up
        2
        ·
        4 months ago

        Your “we” and my “we” are probably not the same, I’m afraid. I’m not shocked that the difference in context would result in a difference of perception, but I’d argue that you guys would need an overhaul on the regulations and safety nets thing regardless.

        • octopus_ink@lemmy.ml
          link
          fedilink
          English
          arrow-up
          3
          ·
          edit-2
          4 months ago

          Fair point, this ol’ nation needs a new set of spark plugs and a valve job, at a minimum. :)

          Edit: DAMMIT how am I a moderator again @[email protected]? Removing myself again now.