• conciselyverbose@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    61
    arrow-down
    3
    ·
    2 months ago

    But in a separate Fortune editorial from earlier this month, Stanford computer science professor and AI expert Fei-Fei Liargued that the “well-meaning” legislation will “have significant unintended consequences, not just for California but for the entire country.”

    The bill’s imposition of liability for the original developer of any modified model will “force developers to pull back and act defensively,” Li argued. This will limit the open-source sharing of AI weights and models, which will have a significant impact on academic research, she wrote.

    Holy shit this is a fucking terrible idea.

    • Zorsith@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      16
      arrow-down
      6
      ·
      2 months ago

      I read that as “incentivizing keeping AI in labs and out of the hands of people who shouldn’t be using it”.

      That said, you’d think they would learn by now from Piracy: once it’s out there, it’s out there. Can’t put it back in the jar.

      • conciselyverbose@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        30
        arrow-down
        1
        ·
        2 months ago

        They should be doing the exact opposite and making it incredibly difficult not to open source it. Major platforms open sourcing much of their systems is basically the only good part of the AI space.

        • Monstrosity@lemm.ee
          link
          fedilink
          English
          arrow-up
          10
          ·
          2 months ago

          Also, they used our general knowledge and culture to train the damn things. They should be open sourced for that reason alone. Llms should be seen and treated like libraries, as collections of our common intellect, accessible by everyone.

      • LainTrain@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        13
        ·
        2 months ago

        Not open-sourcing it is a terrible idea, it just creates more black boxes and gives corporations a further upper hand.

    • ZILtoid1991@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      arrow-down
      1
      ·
      2 months ago

      Same energy as PirateSoftware’s “If AAA companies can’t kill games due to always online DRM then small indie devs have to support their games forever, thus bankrupting them” argument.

    • AbouBenAdhem@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 months ago

      I haven’t yet read Li’s editorial, but I’m generally more inclined to trust her take on these issues than Hinton and Bengio’s.

  • Zarxrax@lemmy.world
    link
    fedilink
    English
    arrow-up
    32
    ·
    2 months ago

    Wtf does a kill switch even mean? PCs have kill switches on them already, in the form of a power switch.

    • Echo Dot@feddit.uk
      link
      fedilink
      English
      arrow-up
      41
      ·
      2 months ago

      I’m afraid the AI has become self-aware and put a piece of tape over the power switch it is now unstoppable.

    • Rentlar@lemmy.ca
      link
      fedilink
      English
      arrow-up
      15
      ·
      2 months ago

      The legislator tried pressing the button on the monitor but the computer kept whirring!!! It’s alive and has a mind of its own!!!

  • Echo Dot@feddit.uk
    link
    fedilink
    English
    arrow-up
    14
    ·
    2 months ago

    You would have assumed that legislators in California of all places would have access to experts that could explain to them why this won’t work.

  • tal@lemmy.today
    link
    fedilink
    English
    arrow-up
    9
    arrow-down
    1
    ·
    2 months ago

    As we’ve previously explored in depth, SB-1047 asks AI model creators to implement a “kill switch” that can be activated if that model starts introducing “novel threats to public safety and security,”

    A model may only be one component of a larger system. Like, there may literally be no way to get unprocessed input through to the model. How can the model creator even do anything about that?