Everyone disliked that.

  • invalidusernamelol [he/him]@hexbear.net
    link
    fedilink
    English
    arrow-up
    21
    ·
    14 days ago

    I think having a policy that forces disclose of LLM code is important. It’s also important to solidify that AI code should only ever be allowed to exist in userland/ring 3. If you can’t hold the author accountable, the code should not have any permissions or be packaged with the OS.

    I can maybe see using an LLM for basic triaging of issues, but I also fear that adding that system will lead to people placing more trust in it than they should have.

      • invalidusernamelol [he/him]@hexbear.net
        link
        fedilink
        English
        arrow-up
        3
        ·
        13 days ago

        I know, that was me just directly voicing that opinion. I do still think that AI code should not be allowed in anything that eve remotely needs security.

        Even if they can still be held accountable, I don’t think it’s a good idea to allow something that is known to hallucinate believable code to write important code. Just makes everything a nightmare to debug.