• Herr_Drosselmeyer@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Oldest trick in the book. Once you’ve established yourself as a market leader, regulations strongly favour you over smaller competitors since you’re better prepared to comply with them, especially if you were involved in shaping them.You’ll gladly trade slightly reduced profit margins for prolonged market dominance.

    Thus, it makes sense to trigger regulations, even if it requires you to embellish the truth.

    • PSMF_Canuck@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      It won’t work. US regulations that don’t make sense will just shift the interesting product development overseas. As has happened in every industry up to now…

      • Herr_Drosselmeyer@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        It’ll work enough for it to be worth it. People don’t like jumping through hoops and often end up just going with what’s easily available. Enthusiasts will download Chinese LLMs but a large American corp? Not going to happen. Truth is, people like us are barely a tiny blip on the radar of commercial enterprises.

    • KillerMiller13@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      They’re willing to sabotage everyone because they won’t be sabotaging themselves that much? That’s messed up.

  • WiSaGaN@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I always thought of Andrew Ng as a more recognizable name than Google Brain.

    • DataAvailability@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Does the argument not make sense? Why not first evaluate the arguments for not open sourcing models on the face instead of reaching for people’s personal incentives to lie about it? Seems like people forgot step one and just went to the assumption of mal intent, like you said you did.

      Given that we don’t really know how AI can be used for malicious purposes, might it make sense that the org with by-far the most powerful model chooses not to release their secrets, as to slow the pace of malicious use?

      Is it possible that Altman believes this, or does his incentive to lie about it so greatly outweigh anything else that you can’t even consider the merits of the argument? I hear way too much about why OAI must be lying about this, not enough considering what they have to say.

  • whyzantium@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    How could they be “lying” about risks? Risks aren’t facts, they’re statements of probability. Big tech companies may benefit from propagating the existence of risk, but they can’t be “lying” about risks unless there’s some scientific study showing that the risk doesn’t exist.

  • psi-love@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    The bad thing is, just because Andrew Ng states this, doesn’t make it true or the possibility of dangers less relevant. There are people outside of big businesses like Hinton who also warn, even though he is not part of “big companies” anymore.

    Also, what is all of this about? In the end there are multiple scenarios in which ways AI can harm society. It probably won’t be about Terminator rising. On the other hand precausions revolve around the fact that we actually don’t really know, because this technology is so new.

    I also don’t think that “big companies” like OpenAI even need to shut down smaller businesses, because - as Sam Altman stated - incoming money really isn’t any issue for them at all. They are drowning in money.

    While there are certainly people who only care for money and other kinds of status symbols, I still believe that many people working within those companies actually try to be truthful about their work as individuals.

  • Zelenskyobama2@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Capitalism = free-market competition

    Socialism = public monopoly (the state, or a private monopoly)

    These AI companies want to put us under a socialist regime. We must stop them.

    • ninjasaid13@alien.topOPB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Socialism = public monopoly (the state, or a private monopoly)

      How is a private monopoly a public monopoly?

      • Zelenskyobama2@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        A company can achieve such high levels of productivity that it no longer relies on the profit motive, whether it’s privately or publicly owned. When I refer to a “private monopoly,” I’m describing a private enterprise that has effectively transitioned into a publicly regulated entity with a focus on improving societal ills.

      • rjames24000@alien.topB
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        reminds me of the DMV … which is a flaming pile of dogshit and treats people like shit because they can do whatever they want and still be in business… you keep your socialism

  • Betaglutamate2@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I honestly think that if there is a chance of AI wiping out humanity we should take it seriously.

    Companies are right now doing the easy thing and saying there is no danger. What I want is proof. Physicists showed proof for example that the LHC was safe.

    If there is even a 0.1% chance of an AI taking over we should be serious about it. Alternatively imagine if I said the following: I am genetically engineering monkeys to be super smart way smarter than humans. I am also going to give them the tools to improve themselves further and to reproduce near instantly then I’m going to release them for 5.99 a month as servants to humans. Hoe many of you would be worried.

    • stonesst@alien.topB
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Some companies are doing the easy thing and saying there’s no danger, and then there’s others that are being honest about the risks they see over the horizon which causes moronic threads like this. There’s no winning… people are too damn cynical

  • Sabin_Stargem@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I am not concerned about the AI itself. Rather, it is about who instills “the rules” into that AI. Are they going to be Asimovian or Robocopic?

    Knowing human history, I am not optimistic.

  • laveshnk@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    Doesn’t basically everyone know who Andrew Ng is? Or at least change google brain to Coursera’s founder maybe…?

  • Naiw80@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    I’m so baffled this has not been realised by people before, it’s so obvious and it’s not the first time in history it happens either.

    First of all Max Tegmark, it’s not even the slightest suspicious that his “non profit” organisation received millions of donations from Elon Musk? I have not figured out what Elons stake in this is yet but I have absolutely no doubt in my mind it’s economical, basically everything he ever did and said has been to manipulate the Stockmarket etc, I doubt that changed recently.

    Then you have OpenAI, that first and formost is everything but Open and very ProprietaryAI nowadays, and what seriously annoys me is that OpenAI in particular been “teasing” about “AGI in n days” etc on several occasions for what purpose if not to manipulate expectations and investors, yet they are one of the most driving in this matter- are people really that stupid that they can’t put together 1 and 1?

  • 218-69@alien.topB
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 year ago

    No shit, they can loophole around any regulation with their gazillions, it’s everyone else that gets fucked.

    Young Man! OpenAI and friends