• onlinepersona@programming.dev
    link
    fedilink
    English
    arrow-up
    28
    ·
    edit-2
    5 hours ago

    Move Godot to codeberg. Github is too popular. It’s where all the fool kids go and link their LinkedIn to. It’s Microslop all the way down. It’s highly likely the number of real contributors will be higher on Codeberg than Github.

    Also, doesn’t github have a viebcoding platform built into it making it really easy to create AI slop?

    Edit: yep they do! Get the fuck off of github!

    • chicken@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      4
      ·
      2 hours ago

      Godot is also weighing the possibility of moving the project to another platform where there might be less incentive for users to “farm” legitimacy as a software developer with AI-generated code contributions. But moving to a less popular platform could run the risk of alienating legitimate contributors.

      • onlinepersona@programming.dev
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 hour ago

        The article does indeed continue 😮 There are so many full page interruptions,I thought the end had been reached.

        Good to know that they are considering moving away from Microslop. It’s about damn time. As the article insinuates, there is little hope of Microslop turning down the dial on AI. They will create AI problems and sell AI solutions for those problems.

  • frongt@lemmy.zip
    link
    fedilink
    English
    arrow-up
    32
    ·
    edit-2
    7 hours ago

    They should close PRs to the public and only accept then from contributors who apply to be vetted.

    It sucks, but that’s really the only good way to prevent spam.

      • addie@feddit.uk
        link
        fedilink
        English
        arrow-up
        5
        ·
        3 hours ago

        AI is quite good at solving captchas; better than many humans. And it doesn’t really slow down the sloppers for them to set their machine running, come back in an hour and then solve a puzzle manually to submit it. Couple of minutes of work every day and they can still drown the world in bullshit.

        Something needs to change, but I’m not convinced that would be enough

  • Lung@lemmy.world
    link
    fedilink
    English
    arrow-up
    18
    ·
    8 hours ago

    Seems like a natural next step is some kinda reputation system for a project’s contributors. If you’ve written 50 successfully merged PRs, you’re certainly less likely to make trash in any method. Create a mentorship heirarchy. It sounds very helpful no matter what. Then the people who have merged 0 PRs to the project will likely work harder too, since they know they will be at the most scrutiny

    • webghost0101@sopuli.xyz
      link
      fedilink
      English
      arrow-up
      8
      ·
      5 hours ago

      Its trivial for ai to successfully merge 50 prs into the projects of other ai.

      It be frustrating for a beginner that every time they try to contribute they are ignored because they have yet to contribute.

  • Lvxferre [he/him]@mander.xyz
    link
    fedilink
    English
    arrow-up
    16
    ·
    8 hours ago

    When the topic of AI submissions flooding open source projects pops up, my immediate reaction is to think "see, this is why you disregard intentions". Because I genuinely believe a lot of the people submitting this slop are trying to help the project, even if in reality they’re harming it, by wasting the maintainers’ time with their crap.They cause harm and deserve to be treated as a source of harm, simple as.

    And while most projects could/should use more money, I don’t think that’s the solution; it allows the devs to handle more workload, sure, but the goal should be to reduce it. I think this will be eventually done through pre-sorting contributors: a cathedral for the outsiders, but a bazaar for the insiders.

      • Lvxferre [he/him]@mander.xyz
        link
        fedilink
        English
        arrow-up
        4
        ·
        3 hours ago

        There’s a lot in your article I agree with. A lot. I could nitpick some of the middle layers, but the conclusion is the same — we should simply disregard intentions, when judging the morality of the actions of someone (incl. ourselves).

        Specially the 7th layer — what you said there is something that has been living in my mind for a long time, but I was never able to phrase it properly.

        About the 8th layer: the bourgeoisie does love to exploit this problem when it helps them to get less blame, since it’s impossible to prove someone doesn’t have good intentions. But I don’t think they created it, I think the problem is older even than our own species, and it comes from developing a theory of mind.

        Thank you for sharing it!

        • Grail@multiverse.soulism.net
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 hours ago

          In the spirit of level 7: I use capitalised pronouns in all three grammatical persons

          I don’t think the Bourgeois created intentionalism, but I do think they’re responsible for its prevalence in our modern society. The degree to which most people are concerned with intentions is so stifling, unhealthy, and unsustainable that I do not believe it can be natural. We live in an age of relatively little social upheaval. There are fewer revolutions and wars than usual right now, and that’s not entirely a good thing. Look at the USA - they are in desperate need of a civil war, or we shall soon be fighting world war 3. Yet the people are far more passive than we would expect from history. Nonviolent resistance is also down, so it’s not just military technology. It’s propaganda.

          I think intentionalism is a meme - an idea that through natural selection has taken on a life of its own. And because it helps the Bourgeoisie, they, perhaps unconsciously, create conditions that favour it in their media empires.

  • Onno (VK6FLAB)@lemmy.radio
    link
    fedilink
    English
    arrow-up
    16
    ·
    9 hours ago

    I wonder if the influx of slop contributions can be stemmed by a legal document that makes the contributor legally liable for their submission.

    Seems like lawyers have been learning the hard way to be accountable for their slop, perhaps these lessons can be leveraged in the open source community.

    • frongt@lemmy.zip
      link
      fedilink
      English
      arrow-up
      9
      ·
      7 hours ago

      Legally liable for what? Just being bad code? How are you going to enforce that against some kid in another country?

    • subignition@fedia.io
      link
      fedilink
      arrow-up
      8
      ·
      8 hours ago

      It’s time to start putting maintainers’ attention behind a paywall. $50 refundable deposit to submit a PR, forfeited if it’s obvious AI slop

    • UnderpantsWeevil@lemmy.world
      link
      fedilink
      English
      arrow-up
      8
      ·
      8 hours ago

      Real “these kids would be very upset if they could read” situation. Who bothers to pick through the whole EULA before submitting?

      Like any open source mass contribution project that’s gained too much popularity, you need extra firebreaks in between the Open Submission and Final Product.

      That means adding manpower to submission review, QA, etc, which public projects don’t often have.

      Sort of the Achilles Heel of the open source community. AI is just making the vulnerability extra glaring.

    • inclementimmigrant@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      6
      ·
      9 hours ago

      That would be a closing a gate after the horses have escaped situation.

      Letting unverifiable code in would damage to developers and users that wouldn’t be easy to disentangle and erode trust in the product, killing it.