• Embargo@lemm.ee
    link
    fedilink
    English
    arrow-up
    29
    arrow-down
    3
    ·
    edit-2
    1 day ago

    Oh no! How will I generate a picture of Sam Altman blowing himself now!?

  • latenightnoir@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    3
    ·
    edit-2
    1 day ago

    Sad to see you leave (not really, tho’), love to watch you go!

    Edit: I bet if any AI developing company would stop acting and being so damned shady and would just ASK FOR PERMISSION, they’d receive a huge amount of data from all over. There are a lot of people who would like to see AGI become a real thing, but not if it’s being developed by greedy and unscrupulous shitheads. As it stands now, I think the only ones who are actually doing it for the R&D and not as eye-candy to glitz away people’s money for aesthetically believable nonsense are a handful of start-up-likes with (not in a condescending way) kids who’ve yet to have their dreams and idealism trampled.

    • daniskarma@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      3
      ·
      21 hours ago

      In Spain we trained an AI using a mix of public resources available for AI training and public resources (legislation, congress sessions, etc). And the AI turned out quite good. Obviously not top of the line, but very good overall.

      It was a public project not a private company.

    • HakFoo@lemmy.sdf.org
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 day ago

      But what data would it be?

      Part of the “gobble all the data” perspective is that you need a broad corpus to be meaningfully useful. Not many people are going to give a $892 billion market cap when your model is a genius about a handful of narrow subjects that you could get deep volunteer support on.

      OTOH maybe there’s probably a sane business in narrow siloed (cheap and efficient and more bounded expectations) AI products: the reinvention of the “expert system” with clear guardrails, the image generator that only does seaside background landscapes but can’t generate a cat to save its life, the LLM that’s a prettified version of a knowledgebase search and NOTHING MORE

      • latenightnoir@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        edit-2
        21 hours ago

        You’ve highlighted exactly why I also fundamentally disagree with the current trend of all things AI being for-profit. This should be 100% non-profit and driven purely by scientific goals, in which case using copyrighted data wouldn’t even be an issue in the first place… It’d be like literally giving someone access to a public library.

        Edit: but to focus on this specific instance, where we have to deal with the here-and-now, I could see them receiving, say, 60-75% of what they have now, hassle-free. At the very least, and uniformly distributed. Again, AI development isn’t what irks most people, it’s calling plagiarism generators and search engine fuck-ups AI and selling them back to the people who generated the databases - or, worse, working toward replacing those people entirely with LLMs! - they used for those abhorrences.

        Train the AI to be factually correct instead and sell it as an easy-to-use knowledge base? Aces! Train the AI to write better code and sell it as an on-board stackoverflow Jr.? Amazing! Even having it as a mini-assistant on your phone so that you have someone to pester you to get the damned laundry out of the washing machine before it starts to stink is a neat thing, but that would require less advertising and shoving down our throats, and more accepting the fact that you can still do that with five taps and a couple of alarm entries.

        Edit 2: oh, and another thing which would require a buttload of humility, but would alleviate a lot of tension would be getting it to cite and link to its sources every time! Have it be transformative enough to give you the gist without shifting into plagiarism, then send you to the source for the details!

  • TheBrideWoreCrimson@sopuli.xyz
    link
    fedilink
    English
    arrow-up
    4
    ·
    16 hours ago

    My main takeaway is that some contrived notion of “national security” has now become an acceptable justification for business decisions in the US.

  • shaggyb@lemmy.world
    link
    fedilink
    English
    arrow-up
    15
    arrow-down
    2
    ·
    22 hours ago

    “How am I supposed to make any money if I can’t steal all of my products to sell back to the world that produced them?”

    Yeah, fuck that. The whole industry deserves to die.

  • Shanmugha@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    13 hours ago

    National security my ass. More like his time span to show more dumb “achievements” while getting richer depends on it and nothing else

  • JHD@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    3
    ·
    23 hours ago

    Strange that no one mentioned openai making money off copyrighted works.

  • SlopppyEngineer@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    ·
    24 hours ago

    He’s afraid of losing his little empire.

    OpenAI also had no clue on recreation the happy little accident that gave them chatGPT3. That’s mostly because their whole thing was using a simple model and brute forcing it with more data, more power, more nodes and then even more data and power until it produced results.

    As expected, this isn’t sustainable. It’s beyond the point of decreasing returns. But Sam here has no idea on how to fix that with much better models so goes back to the one thing he knows: more data needed, just one more terabyte bro, ignore the copyright!

    And now he’s blaming the Chinese into forcing him to use even more data.

  • Greyfoxsolid@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    4 minutes ago

    Sorry to say, but he’s right. For AI to truly flourish in the West, it needs access to all previously human made information and media.

  • glitchdx@lemmy.world
    link
    fedilink
    English
    arrow-up
    48
    arrow-down
    1
    ·
    11 hours ago

    The only way this would be ok is if openai was actually open. make the entire damn thing free and open source, and most of the complaints will go away.

  • Jamdroid@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    14 hours ago

    As far as the ai industry has already broken copyright laws. It will not be actually intelligent for a long time. Just like crypto this seems like a global scam that has squandered resources for a dream of a free workforce. Instead of working together to try and create an ai there are lots of technology companies doing the same ineffective bull 🤔

    • Knock_Knock_Lemmy_In@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      13 hours ago

      Oh yes. Deepseek can quote from copyright sources. So can openAI models, but they are programmed not to.

      Facebook trained on the torrent of Annas archive.

      The copyright horse has left the stable.