Jan Leike, a key safety researcher at firm behind ChatGPT, quit days after launch of its latest AI model, GPT-4o

    • GregorGizeh
      link
      fedilink
      English
      arrow-up
      10
      ·
      edit-2
      6 months ago

      organization is non profit but has a for profit subdivision

      reckless, greedy leader gets ousted by the more careful board

      employees (and investors iirc) throw a fit, leading to the whole more careful faction to get booted from the board

      reckless leader is reinstated and has barely any checks and balances now

      company pursues reckless and greedy strategies

      Surprise?

  • Veraxus@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    edit-2
    6 months ago

    I always narrow my eyes when I hear someone talk about “safety” in the context of AI, because they usually just mean that the AI doesn’t engage in enough moral grandstanding when you ask it sketchy or risqué questions. That’s the same level of pearl-clutching that Tipper Gore espoused over music in the 90s.

    But there are legitimate concerns, like lying about real people and topics, reproducing training data (especially personal information) too closely with the right kind of prompting, etc. The problem is that I can’t tell what kind this person is. Are they upset because the AI can recommend marijuana strains… or because it can do something like leak peoples personal information? The article (and people involved in these efforts) too often lump it all together. See, for example: Anthropic

    Now, all of that said, OpenAI is suuuper creepy. The way they started as a non-profit and then somehow managed to add a for-profit component… that is not acceptable and it’s disgusting that it’s allowed. It makes everything they do suspect and I’m inclined to believe what this exiting researcher says.