Cloudron makes it easy to run web apps like WordPress, Nextcloud, GitLab on your server. Find out more or install now.


Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Bookmarks
  • Search
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

Cloudron Forum

Apps | Demo | Docs | Install
  1. Cloudron Forum
  2. LibreChat
  3. How to use local GPU with remote LibreChat?

How to use local GPU with remote LibreChat?

Scheduled Pinned Locked Moved LibreChat
librechattunnel
5 Posts 3 Posters 61 Views 3 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • L Offline
    L Offline
    LoudLemur
    wrote last edited by
    #1

    Many of us run Cloudron on a remote VPS (Virtual Private Server) without a GPU (Graphical Processing Unit) and then deploy applications like LibreChat there.

    How could we easily make use of our local hardware, which might include a graphics card, to help the inferencing on LibreChat?

    Rathole has been requested for Cloudron but there are other applications which might help, too.

    https://github.com/rapiz1/rathole#rathole

    how about zrok or FRP?

    https://zrok.io/
    https://github.com/fatedier/frp

    marcusquinnM 1 Reply Last reply
    1
    • msproutM Offline
      msproutM Offline
      msprout
      wrote last edited by
      #2

      Imo, just serve Ollama on the server that has the GPU either locally (bare metal) or resource mapped (Proxmox/Virtualized), make sure to pass the server address flag in the systemd module / start script (it's in the docs; sorry, on mobile), connect both machines to a Tailscale tailnet, then configure LibreChat in the two config files to point the Ollama settings to your GPU's tailnet IP or hostname. I have found that this pathway is pretty robust. I haven't noticed any real slowdown, and my VPS and homelab are over 4,000 miles apart. 😊

      1 Reply Last reply
      2
      • msproutM Offline
        msproutM Offline
        msprout
        wrote last edited by
        #3

        Zrok/OpenZiti looks hella cool though.

        If you are committed to self-hosting, you can check out application/protocols like NetBird, HeadScale, innernet, or plain ol' vanilla WireGuard.

        1 Reply Last reply
        2
        • L LoudLemur

          Many of us run Cloudron on a remote VPS (Virtual Private Server) without a GPU (Graphical Processing Unit) and then deploy applications like LibreChat there.

          How could we easily make use of our local hardware, which might include a graphics card, to help the inferencing on LibreChat?

          Rathole has been requested for Cloudron but there are other applications which might help, too.

          https://github.com/rapiz1/rathole#rathole

          how about zrok or FRP?

          https://zrok.io/
          https://github.com/fatedier/frp

          marcusquinnM Offline
          marcusquinnM Offline
          marcusquinn
          wrote last edited by
          #4

          @LoudLemur Have you checked out Kimi.com yet?

          Although I guess you don't have a terrabyte in VRAM, thought you might like this post:

          • https://x.com/iamgrigorev/status/1944488312489570698

          a343276a-861b-4c70-8dcc-6ef90c31d147-image.png

          Web Design https://www.evergreen.je
          Development https://brandlight.org
          Life https://marcusquinn.com

          L 1 Reply Last reply
          2
          • marcusquinnM marcusquinn

            @LoudLemur Have you checked out Kimi.com yet?

            Although I guess you don't have a terrabyte in VRAM, thought you might like this post:

            • https://x.com/iamgrigorev/status/1944488312489570698

            a343276a-861b-4c70-8dcc-6ef90c31d147-image.png

            L Offline
            L Offline
            LoudLemur
            wrote last edited by
            #5

            @marcusquinn Wow! That is amazing. Thank you.

            1 Reply Last reply
            1
            Reply
            • Reply as topic
            Log in to reply
            • Oldest to Newest
            • Newest to Oldest
            • Most Votes


            • Login

            • Don't have an account? Register

            • Login or register to search.
            • First post
              Last post
            0
            • Categories
            • Recent
            • Tags
            • Popular
            • Bookmarks
            • Search