• filister@lemmy.world
    link
    fedilink
    English
    arrow-up
    11
    ·
    2 days ago

    The future is very small models trained to work in a certain domain and able to run on devices.

    Huge foundational models are nice and everything, but they are simply too heavy and expensive to run.

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      1 day ago

      Yeah. You are preaching to the choir here.

      …Still though, I just meant there’s no reason to use Gemma 3 27B (or 12? Whatever they used) unaugmented in AI Studio. The smallest flash seems to be more optimal for TPUs (hence it runs faster).