Im using Ollama on my server with the WebUI. It has no GPU so its not quick to reply but not too slow either.

Im thinking about removing the VM as i just dont use it, are there any good uses or integrations into other apps that might convince me to keep it?

  • Possibly linux@lemmy.zip
    link
    fedilink
    English
    arrow-up
    2
    ·
    3 months ago

    I am really curious if you can get the traditional smart functionality along with a LLM. Maybe have some sort of keyword the prompts the AI. You also could write a custom generated system prompt that includes the weather, time and any other information