• Xylight@lemdro.id
    link
    fedilink
    English
    arrow-up
    2
    ·
    4 days ago

    That is a thing, and it’s called quantization aware training. Some open weight models like Gemma do it.

    The problem is that you need to re-train the whole model for that, and if you also want a full-quality version you need to train a lot more.

    It is still less precise, so it’ll still be worse quality than full precision, but it does reduce the effect.

        • mudkip@lemdro.id
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 days ago

          Is it, or is it not, AI slop? Why are you using so heavily markdown formatting? That is a telltale sign of an LLM being involved

          • psud@aussie.zone
            link
            fedilink
            English
            arrow-up
            1
            ·
            22 hours ago

            heavily markdown formatting

            They used one formatting mark, and it’s the most common. What are you smoking, and may I have some?