• Optional@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 months ago

      But but but . . . They have the money

      Also, and more probably, no one in a decision-making capacity in The Government knows how AI works and really doesn’t want to find out.

  • randon31415@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 months ago

    We will self regulate if you give us a guaranteed monopoly!

    White house: but open source exists.

    Well then no regulation.

  • Asafum@feddit.nl
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 months ago

    I swear the only legitimate response to a corporation saying “I’m going to self regulate”

    Is:

    “HAHAHAHAHAHAHAHA HAHAHAHAHAHAHA!!! Ok anyway, here are the regulations we’re going to enforce.”

    • Alphane Moon@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 months ago

      That fact that is not taken as a given, speaks a lot about how deeply ingrained corruption is in our society.

      I would almost argue that even “neutral” newswires like AP/Reuters should use language like “Companies A, B, C, have created a common PR organisation that will be focused on self-regulation polemics …”.

    • vithigar@lemmy.ca
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 months ago

      Any entity or group that can be trusted to self-regulate wouldn’t raise the question about whether they need to be regulated in the first place.

    • pyre@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 months ago

      yeah how is this even allowed, promise or no promise? why can’t individuals do this? “hey, no need to write any laws on poisoning your boss, promise i will self-regulate.”

  • AutoTL;DR@lemmings.worldB
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 months ago

    This is the best summary I could come up with:


    “We’re grateful for the progress leading companies have made toward fulfilling their voluntary commitments in addition to what is required by the executive order,” says Robyn Patterson, a spokesperson for the White House.

    Without comprehensive federal legislation, the best the US can do right now is to demand that companies follow through on these voluntary commitments, says Brandie Nonnecke, the director of the CITRIS Policy Lab at UC Berkeley.

    After they signed the commitments, Anthropic, Google, Microsoft, and OpenAI founded the Frontier Model Forum, a nonprofit that aims to facilitate discussions and actions on AI safety and responsibility.

    “The natural question is: Does [the technical fix] meaningfully make progress and address the underlying social concerns that motivate why we want to know whether content is machine generated or not?” he adds.

    In the past year, the company has pushed out research on deception, jailbreaking, strategies to mitigate discrimination, and emergent capabilities such as models’ ability to tamper with their own code or engage in persuasion.

    Meanwhile, Microsoft has used satellite imagery and AI to improve responses to wildfires in Maui and map climate-vulnerable populations, which helps researchers expose risks such as food insecurity, forced migration, and disease.


    The original article contains 3,300 words, the summary contains 197 words. Saved 94%. I’m a bot and I’m open source!

    • FiniteBanjo@lemmy.today
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 months ago

      Yes, you’re correct, but technically, the FAA regulated Boeing. But the FAA was desperate for qualified workers, and all of their hires worked for Boeing.

  • j4k3@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    4 months ago

    Any government authoritarian intervention in AI is a joke. The US has a greatly diluted sense of relevance. AI research is international and restrictions will only make the USA irrelevant. Even the names you think of as American are actually funding advancements coming from other countries, most of them in Germany and Asia. I’ve read several white papers on current AI research recently, none were from US based schools or academics. The USA is simply not relevant in some imperial nonsense narrative. AI is a vital military technology. Limiting it through Luddite isolationism will reduce the experience pool massively, and push the research further into places where fresh economic growth is happening without regressive stagnation and corruption of toxic consolidation of wealth.

    • 0x0@programming.dev
      link
      fedilink
      English
      arrow-up
      0
      ·
      4 months ago

      You’re kinda missing the point… the big american names are claiming AI should be limited because it’s so dangerous.

      Who should control AI? Them.

      • j4k3@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        4 months ago
        Only proprietary AI pushes for regulatory measures in an attempt at monopoly. That boils down to one man's corruption.

        Control is impossible. The technology is open source. Transformers are to the present what Apache is to the internet. There were several proprietary attempts to monopolize web servers too, all of them failed because of the open source alternative.

        When one of the old engineers from Bell Labs starts pushing some new open source thing, I pay close attention. The digital age is built on technology with those credentials.

        Yann LeCunn is the head of Meta AI and a primary driver of open source AI. He operates independent of Zuck. Meta AI is not trying to monopolize AI, they are attempting to lead, but not monopolize or control.

        The authoritarian idea of control is a fallacy. This is not something that can be controlled like that. People don’t seem to realize the ultimate scope yet. AI is on par with the entire internet in how it will change society long term.

        In a lot of ways, present AI is like the early days of the microprocessor and personal computers. Most people couldn’t really see the potential uses of computers with an Apple II running a 6502 variant. It was barely more than a tech novelty. The chip itself was pretty useless until all the peripherals were developed around. At present, AI is kinda messy in the public sphere. All of the tools available publicly are like rough examples only. There are a lot more capabilities that are not present in the libraries most people are using in code. This is the disconnect between publicly facing tools and why corporations appear to make foolish decisions. They are being approached by the AI companies directly for integrated solutions. Tangent aside, the current usefulness of AI may look limited to people that can not see the bigger picture, or smaller if you will. At its core, AI can reliably add a new logic element to Boolean math operations; a contextual logic element with flexibility. No matter the issues at present, this is new math. There are countless unexplored places to make use of this revolutionary new math. Asking who should control this is like telling the world about division and then asking who should control division. Any attempt to do so is draconian nonsense. It is a fundamentally bad argument. Division is a phenomenon of the universe. The large language model is an enormous statistical math problem involving word vectors and imaginary numbers. It has packaged human language and culture into a deterministic math problem. Some humans struggle to understand basic division. Most humans struggle to understand vectors and ranked tensors. Of those that understand the latter very few understand the math behind large language models. This does not change the fundamental issue that this is just a math problem that has been publicly shared. The large corporate models do not mean very much. They are fighting to be the best generalist. They all strive to add the same types of safety alignment, but this means nothing. Anyone can do the math and make their own model. The real dangers are not generalist models, it’s specialists. Specialists do not need the enormous datasets.

        The fantasy danger of the machines is nothing more than a modern emergence of a Greek pantheon mythos. At the present “safety” is more about populist stupidity in politics, public ignorance, and creating a way for the average person to interact with a tool where if they understood anything about the real complexity they would write it off as too hard to understand. It has been compromised to dumb it down massively. For instance, I can split up a model and talk to the various underlying entities that underpin alignment and create the various patterns you see in prompt replies. I’m working on understanding the sampler techniques that control this behavior and more. This is done with pytorch in the model loader code and is not related to softmax settings like temperature and token cutoffs.

        Anyways, if AI were taken down from someone essentially turning off the internet, I would not be affected and neither would millions of other people. I can run, code, and train AI completely independently. It is just a complex math problem. Controlling math is as draconian as thought policing.

        In the absolute sense, AGI is still a good ways off. Present AI is not persistent. It has no memory or ability to dynamically change yet. All of those apparent features are done in regular code and fed back into the model with each new prompt.

        In the end, AI requires risk mitigation policies. It is not controllable like that.