• trevor@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    17
    arrow-down
    3
    ·
    3 days ago

    I had zero interest in downloading this shit before because LLMs are just lying slop machines, but if it becomes illegal, I will go out of my way to.

    • BreadstickNinja@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      3 days ago

      I’m with you on principle, but the thing is also like 600 GB of data. Not sure I have the disk space to take a stand on this one.

      • LifeInMultipleChoice@lemmy.dbzer0.com
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 days ago

        There are no lite versions? I was trying to find a small LLM version I can run on an old machine and take it off the internet (or just firewall it) and play around with it to see if there is anything worth learning there for me. I was looking at the lite version of llama but when I tried to run the install on mint I ran into some issues and then had to many drinks to focus on it so I went back to something else. Maybe next weekend. If you have any recommendations I’m all ears

      • trevor@lemmy.blahaj.zone
        link
        fedilink
        English
        arrow-up
        2
        ·
        3 days ago

        No worries. I’ve got some 20TB drives i can throw these on 😁

        Now, running a model that large… Well, I’ll just have to stick with the 8-13b params.

        • Pup Biru@aussie.zone
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 days ago

          you can actually run the model, but it just goes very slowly… i run the 70b model on my m1 mbp, and it technically “requires” 128gb or VRAM - it still runs, just not super fast (though i’d say it’s useable in this case a about 1 word per ~300ms)

    • shirro@aussie.zone
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 days ago

      AI models work really well in some domains. I would absolutely use them for speech to text or text to speech as an assistive technology or use OCR to import docs.