• @jpablo68@infosec.pub
    link
    fedilink
    English
    109 hours ago

    I just want a portable self hosted LLM for specific tasks like programming or language learning.

    • @plixel@programming.dev
      link
      fedilink
      English
      88 hours ago

      You can install Ollama in a docker container and use that to install models to run locally. Some are really small and still pretty effective, like Llama 3.2 is only 3B and some are as little as 1B. It can be accessed through the terminal or you can use something like OpenWeb UI to have a more “ChatGPT” like interface.