• Turun@feddit.de
    link
    fedilink
    arrow-up
    2
    ·
    8 months ago

    Run an open source one. Training requires lots of knowledge and even more hardware resources/time. Fine tuned models are available for free online, there is not much use in training it yourself.

    Options are

    https://github.com/oobabooga/text-generation-webui

    https://github.com/Mozilla-Ocho/llamafile

    https://github.com/ggerganov/llama.cpp

    I recommend llavafiles, as this is the easiest option to run. The GitHub has all the stuff you need in the “quick start” section.

    Though the default is a bit restricted on windows. Since the llavafiles are bundling the LLM weights with the executable and Windows has a 4GB limit on executables you’re restricted to very small models. Workarounds are available though!

    • virku@lemmy.world
      link
      fedilink
      arrow-up
      1
      ·
      8 months ago

      Im gonna give llamafile a go! I want to try to run it at least once with a different set of weights just to see it work and also see different weights handle the same inputs.

      The reason I am asking about training is because of my work where fine tuning our own is going to come knocking soon, so I want to stay a bit ahead of the curve. Even though it already feels like I am late to the party.