Cloudron makes it easy to run web apps like WordPress, Nextcloud, GitLab on your server. Find out more or install now.


Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Bookmarks
  • Search
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

Cloudron Forum

Apps - Status | Demo | Docs | Install
  1. Cloudron Forum
  2. LibreChat
  3. How to use local GPU with remote LibreChat?

How to use local GPU with remote LibreChat?

Scheduled Pinned Locked Moved LibreChat
librechattunnel
5 Posts 3 Posters 1.0k Views 3 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • L Online
    L Online
    LoudLemur
    wrote on last edited by
    #1

    Many of us run Cloudron on a remote VPS (Virtual Private Server) without a GPU (Graphical Processing Unit) and then deploy applications like LibreChat there.

    How could we easily make use of our local hardware, which might include a graphics card, to help the inferencing on LibreChat?

    Rathole has been requested for Cloudron but there are other applications which might help, too.

    https://github.com/rapiz1/rathole#rathole

    how about zrok or FRP?

    https://zrok.io/
    https://github.com/fatedier/frp

    marcusquinnM 1 Reply Last reply
    1
    • msproutM Offline
      msproutM Offline
      msprout
      wrote on last edited by
      #2

      Imo, just serve Ollama on the server that has the GPU either locally (bare metal) or resource mapped (Proxmox/Virtualized), make sure to pass the server address flag in the systemd module / start script (it's in the docs; sorry, on mobile), connect both machines to a Tailscale tailnet, then configure LibreChat in the two config files to point the Ollama settings to your GPU's tailnet IP or hostname. I have found that this pathway is pretty robust. I haven't noticed any real slowdown, and my VPS and homelab are over 4,000 miles apart. 😊

      1 Reply Last reply
      2
      • msproutM Offline
        msproutM Offline
        msprout
        wrote on last edited by
        #3

        Zrok/OpenZiti looks hella cool though.

        If you are committed to self-hosting, you can check out application/protocols like NetBird, HeadScale, innernet, or plain ol' vanilla WireGuard.

        1 Reply Last reply
        2
        • L LoudLemur

          Many of us run Cloudron on a remote VPS (Virtual Private Server) without a GPU (Graphical Processing Unit) and then deploy applications like LibreChat there.

          How could we easily make use of our local hardware, which might include a graphics card, to help the inferencing on LibreChat?

          Rathole has been requested for Cloudron but there are other applications which might help, too.

          https://github.com/rapiz1/rathole#rathole

          how about zrok or FRP?

          https://zrok.io/
          https://github.com/fatedier/frp

          marcusquinnM Offline
          marcusquinnM Offline
          marcusquinn
          wrote on last edited by
          #4

          @LoudLemur Have you checked out Kimi.com yet?

          Although I guess you don't have a terrabyte in VRAM, thought you might like this post:

          • https://x.com/iamgrigorev/status/1944488312489570698

          a343276a-861b-4c70-8dcc-6ef90c31d147-image.png

          Web Design & Development: https://www.evergreen.je
          Technology & Apps: https://www.marcusquinn.com

          L 1 Reply Last reply
          2
          • marcusquinnM marcusquinn

            @LoudLemur Have you checked out Kimi.com yet?

            Although I guess you don't have a terrabyte in VRAM, thought you might like this post:

            • https://x.com/iamgrigorev/status/1944488312489570698

            a343276a-861b-4c70-8dcc-6ef90c31d147-image.png

            L Online
            L Online
            LoudLemur
            wrote on last edited by
            #5

            @marcusquinn Wow! That is amazing. Thank you.

            1 Reply Last reply
            1

            Hello! It looks like you're interested in this conversation, but you don't have an account yet.

            Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.

            With your input, this post could be even better 💗

            Register Login
            Reply
            • Reply as topic
            Log in to reply
            • Oldest to Newest
            • Newest to Oldest
            • Most Votes


            • Login

            • Don't have an account? Register

            • Login or register to search.
            • First post
              Last post
            0
            • Categories
            • Recent
            • Tags
            • Popular
            • Bookmarks
            • Search