I’ve been looking into self-hosting LLMs or stable diffusion models using something like LocalAI and / or Ollama and LibreChat.

Some questions to get a nice discussion going:

  • Any of you have experience with this?
  • What are your motivations?
  • What are you using in terms of hardware?
  • Considerations regarding energy efficiency and associated costs?
  • What about renting a GPU? Privacy implications?
  • fidodo@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    edit-2
    6 months ago

    I tried out ollama. It was trivially easy to set up.

    Stable diffusion is a bit more work, but any power user should be able to figure it out.