• GolfNovemberUniform@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    8 months ago

    More closed and non-customizable systems are much more stable. I guess that’s what GNOME devs are trying to achieve and I don’t really mind it. We have other options for those who need customization. The most used and mainstream one really should be focused on stability. Though I don’t think anyone tried breaking icons before. It’s a bit too much. The app devs will need to make multiple icons for different DEs which is a good thing but shouldn’t be forced like that

    • john89@lemmy.ca
      link
      fedilink
      arrow-up
      1
      ·
      8 months ago

      Ahh, the old “it’s too difficult for us.”

      If the gnome devs are so incompetent, why don’t they just get another job?

    • Björn Tantau@swg-empire.de
      link
      fedilink
      arrow-up
      0
      ·
      8 months ago

      The standards are supposed to be the stable thing. If some part of GNOME advertises itself as following a specific standard then it should remain stable in following that standard.

      • GolfNovemberUniform@lemmy.ml
        link
        fedilink
        arrow-up
        0
        ·
        edit-2
        8 months ago

        That is a misconception that 99% of the devs don’t understand. Sometimes you do need major changes that break stuff to upgrade the base. GNOME started doing it recently. Keeping old bases for a very long time makes them bloated, hacky, slow and unstable

        • LeFantome@programming.dev
          link
          fedilink
          arrow-up
          1
          ·
          8 months ago

          Linus Torvalds is author and maintainer of one of the most successful pieces of software ever written ( software that is decades old and still growing in popularity ).

          What does Linus says about your philosophy that “Sometimes you do need major changes that break stuff to upgrade the base”? I think his first sentence explains where he stands but he expands on his initial point.

          https://lkml.org/lkml/2012/12/23/75

        • LeFantome@programming.dev
          link
          fedilink
          arrow-up
          1
          ·
          8 months ago

          Implementing old standards does not magically result in unstable software. I can create software today that implements decades old standards using whatever whiz-bang tech is in vogue.

          I do lot accept that “old bases” have to succumb to any of the things you suggest either. Refactoring is a thing. You can remove dead code, you can adopt new patterns, you make code modular, you can even extend using new tech if you want.

          Linux is 30 years old ( the basic design is decades older ). Should we throw it out? I vote no but allowing Rust into the kernel seems like a good idea. How old is GCC? How old is Microsoft Office? How old is Firefox? This is software you may use every day. Trust me, your life relies on software that is much, much older. How often do you think they rewrite air traffic control systems or core financial software to to make it more “stable” as you suggest?

          I mostly hear your argument when devs want to try new tech and cannot justify it any other way. Most often the result is something that is far buggier and missing many features. By the time the features return, the new code is at least as bloated as the original. Around then, somebody usually suggests a total rewrite.

          Old architectures are a different story. Sometimes things are not worth fixing in place. In my experience though, this is fairly rare. Even then, in-place migration to something else often makes more sense.

          In my view, if you cannot modernize an old code base, it is a skills issue.