• TheSambassador@lemmy.world
    link
    fedilink
    English
    arrow-up
    4
    ·
    10 months ago

    What reason is there for this when the compiler could just optimize that variable out of existence? This feels like the most hand holdy annoying “feature” unless I’m missing something.

    • frezik@midwest.social
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      10 months ago

      Cleaner code. That’s all.

      If you need to take variable you don’t use for some reason (like it’s a function arg that has to follow an interface, but it doesn’t need a specific parameter in this case), then you can prefix it with an underscore.

      • expr@programming.dev
        link
        fedilink
        arrow-up
        2
        ·
        10 months ago

        That’s what warnings are for and -werror for production builds in literally any other language. This has been a solved problem for a very long time.

        • frezik@midwest.social
          link
          fedilink
          arrow-up
          1
          ·
          10 months ago

          Sure. Tell that to the Go devs.

          If the language weren’t pushed by Google, nobody would pay it any attention. It’s yet another attempt to “do C right” and it makes some odd choices in the attempt.

        • dbx12@programming.dev
          link
          fedilink
          arrow-up
          1
          arrow-down
          1
          ·
          10 months ago

          I for my part prefer it that way. Makes sure the code stays clean and nobody can just silence the warnings and be done with it. Because why would you accept useless variables that clutter the code in production builds? Imagine coming back after some time and try to understand the code again. At least you have the guarantee the variable is used somehow and not just “hmm, what does this do? … ah, it’s unused”

          • expr@programming.dev
            link
            fedilink
            arrow-up
            1
            ·
            10 months ago

            …you don’t accept them. Basically every programming language accepts some kind of -werror flag to turn warnings into errors. Warnings for development builds, errors for production builds. This has been a solved problem for a very long time. Not only is it assinine to force them to be errors always, it’s semantically incorrect. Errors should be things that prevent the code from functioning in some capacity.

            • dbx12@programming.dev
              link
              fedilink
              arrow-up
              1
              ·
              10 months ago

              Oh, that makes warnings errors and does not mean “ignore errors”. I’m not too familiar with compiler flags. You could do some mental gymnastics to argue that the unused variable causes the compiler to exit and thus the code is not functioning and thus the unused variable is not a warning but an error :^)

              • expr@programming.dev
                link
                fedilink
                arrow-up
                1
                ·
                10 months ago

                It’s a pretty standard flag in basically all compiled languages, just goes by a different name. -werror in C, -Werror in Java, TreatWarningsAsErrors in C#, etc.