The Pentagon has its eye on the leading AI company, which this week softened its ban on military use.

  • funkforager@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    3
    ·
    8 months ago

    Remember when open ai was a nonprofit first and foremost, and we were supposed to trust they would make AI for good and not evil? Feels like it was only Thanksgiving…

    • Dave@lemmy.nz
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 months ago

      I mean, there was all that drama where the board formed to prevent this from happening kicked out the CEO trying to do this stuff, then the board got booted out and replaced with a new board and brought back that CEO guy. So this was pretty much going to happen.

      • hoshikarakitaridia@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        ·
        8 months ago

        And some people pointed it out even back then. There were signs that the employees were very loyal to Altmann, but Altmann didn’t meet the security concerns of the board. So stuff like this was just a matter of time.

    • Moira_Mayhem@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 months ago

      It seems to be a trend that any service that claims not to be evil is just waiting for the right moment to drop that pretense.

  • Fedizen@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 months ago

    I can’t wait until we find out AI trained on military secrets is leaking military secrets.

  • annehathway12@kbin.social
    link
    fedilink
    arrow-up
    1
    ·
    edit-2
    5 months ago

    It’s interesting to note OpenAI’s decision regarding the ban on using ChatGPT for “Military and Warfare” applications. For more updates and insights on AI developments, visit ChatGPT.

  • kromem@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    8 months ago

    Literally no one is reading the article.

    The terms still prohibit use to cause harm.

    The change is that a general ban on military use has been removed in favor of a generalized ban on harm.

    So for example, the Army could use it to do their accounting, but not to generate a disinformation campaign against a hostile nation.

    If anyone actually really read the article, we could have a productive conversation around whether any military usage is truly harmless, the nuances of the usefulness of a military ban in a world where so much military labor is outsourced to private corporations which could ‘launder’ terms compliance, or the general inability of terms to preemptively prevent harmful use at all.

    Instead, we have people taking the headline only and discussing AI being put in charge of nukes.

    Lemmy seems to care a lot more about debating straw men arguments about how terrible AI is than engaging with reality.

  • mechoman444@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 months ago

    If you guys think that AI hasn’t already been in use in various militarys including America y’all are living in lala land.

    • kromem@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 months ago

      That would count as harm and be disallowed by the current policy.

      But a military application of using GPT to identify and filter misinformation would not be harm, and would have been prevented by the previous policy prohibiting any military use, but would be allowed under the current policy.

      Of course, it gets murkier if the military application of identifying misinformation later ends up with a drone strike on the misinformer. In theory they could submit a usage description of “identify misinformation” which appears to do no harm, but then take the identifications to cause harm.

      Which is part of why a broad ban on military use may have been more prudent than a ban only on harmful military usage.