Cloudron makes it easy to run web apps like WordPress, Nextcloud, GitLab on your server. Find out more or install now.


Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Bookmarks
  • Search
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

Cloudron Forum

Apps - Status | Demo | Docs | Install
  1. Cloudron Forum
  2. Support
  3. Error in apps accessing NFS volume after upgrade to v9.1.6

Error in apps accessing NFS volume after upgrade to v9.1.6

Scheduled Pinned Locked Moved Solved Support
nfsnetwork
33 Posts 2 Posters 62 Views 2 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • S Online
    S Online
    Stirchley.coop
    wrote last edited by james
    #1

    Hi, our Cloudron instance automatically upgraded from v.9.1.5 to v9.1.6 last night. Today many of our apps are in the Error state, showing the following error message:

    BoxError: Storage volume "NetCup RS voln639290a1" is not active. Could not determine mount failure reason. 
    

    All these apps have their data storage located on that volume. I have remounted the volume and restarted the server and individual apps multiple times, with no effect.

    On logging onto the server I found that nfs-server service was not running (did not exist), and the nfs-kernel-server apt package was not installed. This is strange as the NFS mount has been fine for many months/years, and how would it be running without that service? Had Cloudron uninstalled it? I might be wrong about that, I'm not great with Linux but I installed the nfs-kernel-server:

    root@-REDACTED-:~# sudo apt list nfs-*
    Listing... Done
    nfs-common/jammy-updates,now 1:2.6.1-1ubuntu1.2 amd64 [installed]
    nfs-ganesha-ceph/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha-doc/jammy-updates 3.5-1ubuntu1 all
    nfs-ganesha-gluster/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha-gpfs/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha-mem/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha-mount-9p/jammy-updates 3.5-1ubuntu1 all
    nfs-ganesha-nullfs/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha-proxy/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha-rados-grace/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha-rgw/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha-vfs/jammy-updates 3.5-1ubuntu1 amd64
    nfs-ganesha/jammy-updates 3.5-1ubuntu1 amd64
    nfs-kernel-server/jammy-updates,now 1:2.6.1-1ubuntu1.2 amd64 [installed]
    

    The nfs-server service is now running:

    root@-REDACTED-:~# sudo systemctl status nfs-server
    ● nfs-server.service - NFS server and services
         Loaded: loaded (/lib/systemd/system/nfs-server.service; enabled; vendor preset: enabled)
         Active: active (exited) since Fri 2026-04-24 12:57:35 UTC; 3h 53min ago
        Process: 887 ExecStartPre=/usr/sbin/exportfs -r (code=exited, status=0/SUCCESS)
        Process: 888 ExecStart=/usr/sbin/rpc.nfsd (code=exited, status=0/SUCCESS)
       Main PID: 888 (code=exited, status=0/SUCCESS)
            CPU: 11ms
    
    Apr 24 12:57:35 -REDACTED- systemd[1]: Starting NFS server and services...
    Apr 24 12:57:35 -REDACTED-systemd[1]: Finished NFS server and services.
    

    But the above issue still remains after rebooting the server.

    Here is the output of cloudron-support --troubleshoot:

    root@-REDACTED-:~# cloudron-support --troubleshoot
    Vendor: netcup Product: KVM Server
    Linux: 5.15.0-176-generic
    Ubuntu: jammy 22.04
    Execution environment: kvm
    Processor: AMD EPYC 9634 84-Core Processor x 8
    RAM: 16371020KB
    Disk: /dev/vda3       130G
    [OK]    node version is correct
    [OK]    IPv6 is enabled and public IPv6 address is working
    [OK]    docker is running
    [OK]    docker version is correct
    [OK]    MySQL is running
    [OK]    netplan is good
    [OK]    DNS is resolving via systemd-resolved
    [OK]    unbound is running
    [OK]    nginx is running
    [OK]    dashboard cert is valid
    [OK]    dashboard is reachable via loopback
    [OK]    No pending database migrations
    [OK]    Service 'mysql' is running and healthy
    [OK]    Service 'postgresql' is running and healthy
    [OK]    Service 'mongodb' is running and healthy
    [FAIL]  Service 'mail' container is not running (state: created)!
    [OK]    Service 'graphite' is running and healthy
    [OK]    Service 'sftp' is running and healthy
    [OK]    box v9.1.6 is running
    [OK]    Dashboard is reachable via domain name
            Domain stirchley.coop expiry check skipped because whois is not installed. Run 'apt install whois' to check
    

    I'm seeing these errors in /home/yellowtent/platformdata/logs/box.log:

    2026-04-23T21:03:00.049Z shell: mounts: mountpoint -q -- /mnt/volumes/00fd99454f384c91a83326ad3097bb1e errored BoxError: mountpoint exited with code 32 signal null
        at ChildProcess.<anonymous> (file:///home/yellowtent/box/src/shell.js:70:23)
        at ChildProcess.emit (node:events:508:28)
        at maybeClose (node:internal/child_process:1101:16)
        at Socket.<anonymous> (node:internal/child_process:457:11)
        at Socket.emit (node:events:508:28)
        at Pipe.<anonymous> (node:net:346:12) {
      reason: 'Shell Error',
      details: {},
      stdout: '',
      stdoutString: '',
      stdoutLineCount: 0,
      stderr: '',
      stderrString: '',
      stderrLineCount: 0,
      code: 32,
      signal: null,
      timedOut: false,
      terminated: false
    }
    2026-04-23T21:03:00.050Z shell: mounts: systemd-escape -p --suffix=mount /mnt/volumes/00fd99454f384c91a83326ad3097bb1e
    2026-04-23T21:03:00.062Z shell: mounts: journalctl -u mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount\n -n 10 --no-pager -o json
    2026-04-23T21:03:00.074Z apphealthmonitor: app health: 6 running / 5 stopped / 7 unresponsive
    2026-04-23T21:03:00.084Z scheduler: BoxError: Storage volume "NetCup RS voln639290a1" is not active. Could not determine mount failure reason.
        at getAddonMounts (file:///home/yellowtent/box/src/docker.js:247:54)
        at process.processTicksAndRejections (node:internal/process/task_queues:103:5)
        at async getMounts (file:///home/yellowtent/box/src/docker.js:297:25)
        at async Object.createSubcontainer (file:///home/yellowtent/box/src/docker.js:649:20) {
      reason: 'Bad State',
      details: {}
    }
    ...
    2026-04-24T12:31:28.345Z shell: mounts: /usr/bin/sudo --non-interactive /home/yellowtent/box/src/scripts/remountmount.sh /mnt/volumes/00fd99454f384c91a83326ad3097bb1e
    2026-04-24T12:31:28.397Z shell: mounts: /usr/bin/sudo --non-interactive /home/yellowtent/box/src/scripts/remountmount.sh /mnt/volumes/00fd99454f384c91a83326ad3097bb1e errored BoxError: /usr/bin/sudo exited with code 1 signal null
        at ChildProcess.<anonymous> (file:///home/yellowtent/box/src/shell.js:70:23)
        at ChildProcess.emit (node:events:508:28)
        at maybeClose (node:internal/child_process:1101:16)
        at ChildProcess._handle.onexit (node:internal/child_process:305:5) {
      reason: 'Shell Error',
      details: {},
      stdout: '',
      stdoutString: '',
      stdoutLineCount: 0,
      stderr: 'Job failed. See "journalctl -xe" for details.\n',
      stderrString: 'Job failed. See "journalctl -xe" for details.\n',
      stderrLineCount: 1,
      code: 1,
      signal: null,
      timedOut: false,
      terminated: false
    }
    2026-04-24T12:31:28.454Z shell: mounts: mountpoint -q -- /mnt/volumes/00fd99454f384c91a83326ad3097bb1e
    ...
    2026-04-24T12:58:00.215Z scheduler: BoxError: (HTTP code 400) bad parameter - No such container: null
        at Object.createSubcontainer (file:///home/yellowtent/box/src/docker.js:770:28)
        at process.processTicksAndRejections (node:internal/process/task_queues:103:5) {
      reason: 'Docker Error',
      details: {},
      nestedError: Error: (HTTP code 400) bad parameter - No such container: null
          at /home/yellowtent/box/node_modules/docker-modem/lib/modem.js:383:17
          at getCause (/home/yellowtent/box/node_modules/docker-modem/lib/modem.js:418:7)
          at Modem.buildPayload (/home/yellowtent/box/node_modules/docker-modem/lib/modem.js:379:5)
          at IncomingMessage.<anonymous> (/home/yellowtent/box/node_modules/docker-modem/lib/modem.js:347:16)
          at IncomingMessage.emit (node:events:520:35)
          at endReadableNT (node:internal/streams/readable:1701:12)
          at process.processTicksAndRejections (node:internal/process/task_queues:89:21) {
        reason: 'bad parameter',
        statusCode: 400,
        json: { message: 'No such container: null' }
      }
    }
    

    This was the output for journalctl -xe when remounting the volume:

    Apr 24 16:35:55 -REDACTED-systemd[1]: var-lib-docker-overlay2-2e77e0df312b7a38d974f58ce03709aeb8f5b83a3a95b6560a74c107d7e68417-merged.mount: Deactivated success>
    ░░ Subject: Unit succeeded
    ░░ Defined-By: systemd
    ░░ Support: http://www.ubuntu.com/support
    ░░
    ░░ The unit var-lib-docker-overlay2-2e77e0df312b7a38d974f58ce03709aeb8f5b83a3a95b6560a74c107d7e68417-merged.mount has successfully entered the 'dead' state.
    Apr 24 16:36:01 -REDACTED-kernel: Packet dropped: IN=eth0 OUT= MAC=6a:80:31:09:cf:8d:84:03:28:62:58:1a:08:00 SRC=185.242.226.71 DST=152.53.33.202 LEN=40 TOS=0x0>
    Apr 24 16:36:22 -REDACTED-sshd[29055]: Connection closed by authenticating user -REDACTED- 77.90.185.40 port 46604 [preauth]
    Apr 24 16:36:25 -REDACTED-kernel: Packet dropped: IN=eth0 OUT= MAC=6a:80:31:09:cf:8d:84:03:28:62:58:1a:08:00 SRC=51.148.95.245 DST=152.53.33.202 LEN=60 TOS=0x00>
    Apr 24 16:36:57 -REDACTED-kernel: Packet dropped: IN=eth0 OUT= MAC=6a:80:31:09:cf:8d:84:03:28:62:58:1a:08:00 SRC=51.148.95.245 DST=152.53.33.202 LEN=60 TOS=0x00>
    Apr 24 16:37:04 -REDACTED-sudo[29078]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=998)
    Apr 24 16:37:04 -REDACTED-systemd[1]: Unmounting NetCup RS voln639290a1...
    ░░ Subject: A stop job for unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has begun execution
    ░░ Defined-By: systemd
    ░░ Support: http://www.ubuntu.com/support
    ░░
    ░░ A stop job for unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has begun execution.
    ░░
    ░░ The job identifier is 1013.
    Apr 24 16:37:04 -REDACTED-systemd[1]: mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount: Deactivated successfully.
    ░░ Subject: Unit succeeded
    ░░ Defined-By: systemd
    ░░ Support: http://www.ubuntu.com/support
    ░░
    ░░ The unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has successfully entered the 'dead' state.
    Apr 24 16:37:04 -REDACTED-systemd[1]: Unmounted NetCup RS voln639290a1.
    ░░ Subject: A stop job for unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has finished
    ░░ Defined-By: systemd
    ░░ Support: http://www.ubuntu.com/support
    ░░
    ░░ A stop job for unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has finished.
    ░░
    ░░ The job identifier is 1013 and the job result is done.
    Apr 24 16:37:04 -REDACTED-systemd[1]: mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount: Directory /mnt/volumes/00fd99454f384c91a83326ad3097bb1e to mount over >
    ░░ Subject: Mount point is not empty
    ░░ Defined-By: systemd
    ░░ Support: http://www.ubuntu.com/support
    ░░
    ░░ The directory /mnt/volumes/00fd99454f384c91a83326ad3097bb1e is specified as the mount point (second field in
    ░░ /etc/fstab or Where= field in systemd unit file) and is not empty.
    ░░ This does not interfere with mounting, but the pre-exisiting files in
    ░░ this directory become inaccessible. To see those over-mounted files,
    ░░ please manually mount the underlying file system to a secondary
    ░░ location.
    Apr 24 16:37:04 -REDACTED-systemd[1]: Mounting NetCup RS voln639290a1...
    ░░ Subject: A start job for unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has begun execution
    ░░ Defined-By: systemd
    ░░ Support: http://www.ubuntu.com/support
    ░░
    ░░ A start job for unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has begun execution.
    ░░
    ░░ The job identifier is 1014.
    Apr 24 16:37:04 -REDACTED-systemd[1]: Mounted NetCup RS voln639290a1.
    ░░ Subject: A start job for unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has finished successfully
    ░░ Defined-By: systemd
    ░░ Support: http://www.ubuntu.com/support
    ░░
    ░░ A start job for unit mnt-volumes-00fd99454f384c91a83326ad3097bb1e.mount has finished successfully.
    ░░
    ░░ The job identifier is 1014.
    Apr 24 16:37:04 -REDACTED-sudo[29078]: pam_unix(sudo:session): session closed for user root
    Apr 24 16:37:23 -REDACTED-kernel: Packet dropped: IN=eth0 OUT= MAC=6a:80:31:09:cf:8d:84:03:28:62:58:1a:08:00 SRC=88.210.63.192 DST=152.53.33.202 LEN=40 TOS=0x00>
    

    Here are the logs from one of the erroring apps:

    Apr 24 13:57:12 13:M 24 Apr 2026 12:57:12.786 # Redis is now ready to exit, bye bye...
    Apr 24 13:57:12 13:M 24 Apr 2026 12:57:12.786 * DB saved on disk
    Apr 24 13:57:12 13:M 24 Apr 2026 12:57:12.786 * Removing the pid file.
    Apr 24 13:57:12 13:signal-handler (1777035432) Received SIGTERM scheduling shutdown...
    Apr 24 13:57:12 2026-04-24 12:57:12,759 INFO waiting for redis, redis-service to die
    Apr 24 13:57:12 2026-04-24 12:57:12,759 WARN received SIGTERM indicating exit request
    Apr 24 13:57:12 2026-04-24 12:57:12,762 WARN stopped: redis-service (terminated by SIGTERM)
    Apr 24 13:57:12 2026-04-24 12:57:12,788 INFO stopped: redis (exit status 0)
    Apr 24 13:57:37 2026-04-24 12:57:37,995 INFO Included extra file "/etc/supervisor/conf.d/redis-service.conf" during parsing
    Apr 24 13:57:37 2026-04-24 12:57:37,996 INFO Included extra file "/etc/supervisor/conf.d/redis.conf" during parsing
    Apr 24 13:57:37 2026-04-24 12:57:37,996 INFO Set uid to user 0 succeeded
    Apr 24 13:57:37 ==> Starting supervisor
    Apr 24 13:57:38 2026-04-24 12:57:38,002 CRIT Server 'inet_http_server' running without any HTTP authentication checking
    Apr 24 13:57:38 2026-04-24 12:57:38,002 INFO RPC interface 'supervisor' initialized
    Apr 24 13:57:38 2026-04-24 12:57:38,002 INFO supervisord started with pid 1
    Apr 24 13:57:39 13:C 24 Apr 2026 12:57:39.029 # WARNING Memory overcommit must be enabled! Without it, a background save or replication may fail under low memory condition. Being disabled, it can also cause failures without low memory condition, see https://github.com/jemalloc/jemalloc/issues/1328. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect.
    Apr 24 13:57:39 13:C 24 Apr 2026 12:57:39.029 * Configuration loaded
    Apr 24 13:57:39 13:C 24 Apr 2026 12:57:39.029 * Redis version=8.4.0, bits=64, commit=00000000, modified=1, pid=13, just started
    Apr 24 13:57:39 13:C 24 Apr 2026 12:57:39.029 * oO0OoO0OoO0Oo Redis is starting oO0OoO0OoO0Oo
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.030 * Increased maximum number of open files to 10032 (it was originally set to 1024).
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.030 * monotonic clock: POSIX clock_gettime
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.032 # Failed to write PID file: Permission denied
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.032 * Running mode=standalone, port=6379.
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.033 * Server initialized
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.036 * Loading RDB produced by version 8.4.0
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.036 * RDB age 27 seconds
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.036 * RDB memory usage when created 4.76 Mb
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.086 * Done loading RDB, keys loaded: 175, keys expired: 0.
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.087 * DB loaded from disk: 0.053 seconds
    Apr 24 13:57:39 13:M 24 Apr 2026 12:57:39.087 * Ready to accept connections tcp
    Apr 24 13:57:39 2026-04-24 12:57:39,004 INFO spawned: 'redis' with pid 13
    Apr 24 13:57:39 2026-04-24 12:57:39,006 INFO spawned: 'redis-service' with pid 14
    Apr 24 13:57:39 Redis service endpoint listening on http://:::3000
    Apr 24 13:57:40 2026-04-24 12:57:40,639 INFO success: redis entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
    Apr 24 13:57:40 2026-04-24 12:57:40,639 INFO success: redis-service entered RUNNING state, process has stayed up for > than 1 seconds (startsecs)
    Apr 24 14:01:38 [GET] /healthcheck
    Apr 24 14:01:38 BoxError: Storage volume "NetCup RS voln639290a1" is not active. Could not determine mount failure reason.
    Apr 24 14:01:38 }
    Apr 24 14:01:38 }
    

    The volume in the Cloudron /volumes view looks like it has mounted successfully. It has a green circle next to it.
    3753f033-9cfd-4750-b181-30190e429bd0-image.jpeg

    Clicking on File manager allows me to browse the files within it:
    b340954d-4126-4734-b728-c481c322859a-image.jpeg

    I'm out of ideas, and searching online isn't turning up anything helpful I can see. Is it possible the update has done something to our NFS settings or has there been a bug introduced where apps with their data hosted on an NFS volume don't work? This issue has come at a critical time for our organisation so would really appreciate any help you can give!

    Thanks,
    Stu

    1 Reply Last reply
    1
    • jamesJ Online
      jamesJ Online
      james
      Staff
      wrote last edited by james
      #2

      Hello @stirchley.coop and welcome to the Cloudron forum

      Could you please check the output of the following command from your server?

      uname -r
      

      and post the result.
      We have found a Linux Kernel regression of Linux Kernel version 110 which resulted in many networking issues due to the iptables working in new way compared to previous versions.
      Cloudron version v9.1.7 has been published as 'unstable' which fixes this issue with the iptables.
      Since NFS is very much network related, this could also be your issue.
      So you could try to update to Cloudron version 9.1.7 or apply the iptables patch manually.
      If you'd like to try, apply this patch manually https://git.cloudron.io/platform/box/-/commit/bffc97ab2fedb06100bc507abd0f8b42d343ea6e and then systemctl restart cloudron-firewall, but better to simply update Cloudron to version 9.1.7.

      Cloudron does not uninstall any apt packages.
      Also, nfs-kernel-server is not part of the default Cloudron installation.
      Thus, support for custom changes made to your Cloudron is tricky at best.

      1 Reply Last reply
      0
      • S Online
        S Online
        Stirchley.coop
        wrote last edited by
        #3

        Hi here are the results:

        root@v2202403157799261297:~# uname -r
        5.15.0-176-generic
        

        I only tried installing nfs-kernel-server as it was suggested as a fix for NFS not working, so I might just uninstall it if it doesn't affect anything.

        I will try updating now. Thanks for your help!

        jamesJ 1 Reply Last reply
        1
        • S Online
          S Online
          Stirchley.coop
          wrote last edited by
          #4

          I updated to 9.1.7 and now there is a big error at the top of the page:

          646d7e5a-9403-40b0-b678-78890cd09561-image.jpeg

          Error message:

          Platform startup failed
          Failed to start services. f472efd072b2282e1479f0dd3a3bbd6e75e7a1f82d6d65e66071c370c5b48d4f docker: Error response from daemon: failed to set up container networking: driver failed programming external connectivity on endpoint mail (389b0e33027509943a51f0908370f4bcb89cf1add91a68aee6e8ddbfd50b3590): failed to bind host port 0.0.0.0:995/tcp: address already in use Run 'docker run --help' for more information
          
          1 Reply Last reply
          0
          • S Online
            S Online
            Stirchley.coop
            wrote last edited by
            #5

            Is there a way to downgrade to 9.1.6? I can't see an option to do this from the UI

            jamesJ 1 Reply Last reply
            0
            • S Stirchley.coop

              Hi here are the results:

              root@v2202403157799261297:~# uname -r
              5.15.0-176-generic
              

              I only tried installing nfs-kernel-server as it was suggested as a fix for NFS not working, so I might just uninstall it if it doesn't affect anything.

              I will try updating now. Thanks for your help!

              jamesJ Online
              jamesJ Online
              james
              Staff
              wrote last edited by james
              #6

              Hello @stirchley.coop

              @Stirchley.coop said:

              I only tried installing nfs-kernel-server as it was suggested as a fix for NFS not working, so I might just uninstall it if it doesn't affect anything.

              As a rule of thumb for Cloudron.
              Cloudron should work out of the box without any issues.
              Any suggested fix with installing apt packages should be considered to cause more harm than solve anything.
              I would advise to always open a topic in our forum first before trying anything on your own.


              For the port 995 already in use issue.

              Could it be that by installation of some apt package, some recommended other apt packages got installed which is now using port 995?

              Can you check the following command:

              netstat -tulpen | grep -i 995
              

              And also:

              lsof -i :995
              

              Please share the output of these commands.

              1 Reply Last reply
              0
              • S Stirchley.coop

                Is there a way to downgrade to 9.1.6? I can't see an option to do this from the UI

                jamesJ Online
                jamesJ Online
                james
                Staff
                wrote last edited by
                #7

                Hello @stirchley.coop

                @Stirchley.coop said:

                Is there a way to downgrade to 9.1.6? I can't see an option to do this from the UI

                No. Downgrading Cloudron is not an option.

                1 Reply Last reply
                0
                • S Online
                  S Online
                  Stirchley.coop
                  wrote last edited by
                  #8

                  Hi James.
                  Output:

                  root@v2202403157799261297:~# netstat -tulpen | grep -i 995
                  Command 'netstat' not found, but can be installed with:
                  apt install net-tools
                  root@v2202403157799261297:~# lsof -i :995
                  root@v2202403157799261297:~#
                  

                  Should I apt install net-tools?

                  1 Reply Last reply
                  1
                  • jamesJ Online
                    jamesJ Online
                    james
                    Staff
                    wrote last edited by
                    #9

                    Hello @stirchley.coop
                    Yes you can install net-tools.

                    1 Reply Last reply
                    0
                    • S Online
                      S Online
                      Stirchley.coop
                      wrote last edited by
                      #10

                      Ok after installing I get this:

                      root@v2202403157799261297:~# netstat -tulpen | grep -i 995
                      root@v2202403157799261297:~# lsof -i :995
                      root@v2202403157799261297:~#
                      
                      1 Reply Last reply
                      0
                      • S Online
                        S Online
                        Stirchley.coop
                        wrote last edited by
                        #11

                        I'm thinking I uninstall Cloudron from the server and reinstall version 9.1.5 with yesterday's backup

                        jamesJ 1 Reply Last reply
                        1
                        • jamesJ Online
                          jamesJ Online
                          james
                          Staff
                          wrote last edited by
                          #12

                          Hello @stirchley.coop
                          Can you please share the output of the command:

                          docker ps -a
                          

                          and

                          docker network list
                          
                          1 Reply Last reply
                          0
                          • S Online
                            S Online
                            Stirchley.coop
                            wrote last edited by Stirchley.coop
                            #13

                            @james

                            root@v2202403157799261297:~# docker ps -a
                            CONTAINER ID   IMAGE                                      COMMAND                CREATED          STATUS    PORTS     NAMES
                            f472efd072b2   registry.docker.com/cloudron/mail:3.18.2   "/app/code/start.sh"   17 minutes ago   Created             mail
                            root@v2202403157799261297:~# docker network list
                            NETWORK ID     NAME       DRIVER    SCOPE
                            cbe520222cfe   bridge     bridge    local
                            e28b8f24a89c   cloudron   bridge    local
                            f8d98b330cd7   host       host      local
                            d58248fbb326   none       null      local
                            root@v2202403157799261297:~#
                            
                            1 Reply Last reply
                            0
                            • S Stirchley.coop

                              I'm thinking I uninstall Cloudron from the server and reinstall version 9.1.5 with yesterday's backup

                              jamesJ Online
                              jamesJ Online
                              james
                              Staff
                              wrote last edited by
                              #14

                              Hello @stirchley.coop

                              @Stirchley.coop said:

                              I'm thinking I uninstall Cloudron from the server and reinstall version 9.1.5 with yesterday's backup

                              This would be the easy / safe route.
                              I think we can figure out this issue together and fix your server without restoring from a backup.
                              But the decision is up to you.


                              This output:

                              root@v2202403157799261297:~# docker ps -a
                              CONTAINER ID   IMAGE                                      COMMAND                CREATED          STATUS    PORTS     NAMES
                              f472efd072b2   registry.docker.com/cloudron/mail:3.18.2   "/app/code/start.sh"   17 minutes ago   Created             mail
                              

                              is very suspicious.
                              No other docker container running from cloudron except this one.

                              Can you try to restart the box.service with:

                              systemcrtl restart box.service
                              

                              Then you can also observe the box.log with:

                              tail -f /home/yellowtent/platformdata/logs/box.log
                              

                              To see all log messages from Cloudron.

                              1 Reply Last reply
                              0
                              • S Online
                                S Online
                                Stirchley.coop
                                wrote last edited by james
                                #15

                                @james a lot of network errors so far:

                                root@v2202403157799261297:~# systemcrtl restart box.service
                                Command 'systemcrtl' not found, did you mean:
                                  command 'systemctl' from deb systemd (249.11-0ubuntu3.20)
                                  command 'systemctl' from deb systemctl (1.4.4181-1.1)
                                Try: apt install <deb name>
                                root@v2202403157799261297:~# systemctl restart box.service
                                root@v2202403157799261297:~# tail -f /home/yellowtent/platformdata/logs/box.log
                                2026-04-24T17:54:34.309Z docker: deleteContainer: deleting turn
                                2026-04-24T17:54:34.310Z services: startTurn: starting turn container
                                2026-04-24T17:54:34.310Z shell: services: /bin/bash -c docker run --restart=unless-stopped -d --name=turn                 --hostname turn                 --net host                 --log-driver syslog                 --log-opt syslog-address=unix:///home/yellowtent/platformdata/logs/syslog.sock                 --log-opt syslog-format=rfc5424                 --log-opt tag=turn                 -m 268435456                 --memory-swap -1                 -e CLOUDRON_TURN_SECRET=-REDACTED- -e CLOUDRON_REALM=my.stirchley.coop                 -e CLOUDRON_VERBOSE_LOGS=                 --label isCloudronManaged=true                 --read-only -v /tmp -v /run registry.docker.com/cloudron/turn:1.8.2@sha256:9f3609969a5757837505c584c98246a3035a84a273b9be491665ac026423fd5f
                                2026-04-24T17:54:34.526Z services: startMysql: stopping and deleting previous mysql container
                                2026-04-24T17:54:34.526Z docker: stopContainer: stopping container mysql
                                2026-04-24T17:54:34.529Z docker: deleteContainer: deleting mysql
                                2026-04-24T17:54:34.530Z services: startMysql: starting mysql container
                                2026-04-24T17:54:34.530Z shell: services: /bin/bash -c docker run --restart=unless-stopped -d --name=mysql                 --hostname mysql                 --net cloudron                 --net-alias mysql                 --log-driver syslog                 --log-opt syslog-address=unix:///home/yellowtent/platformdata/logs/syslog.sock                 --log-opt syslog-format=rfc5424                 --log-opt tag=mysql                 --ip xxx                 -e CLOUDRON_MYSQL_TOKEN=xxx                 -e CLOUDRON_MYSQL_ROOT_HOST=xxx                 -e CLOUDRON_MYSQL_ROOT_PASSWORD=xxx -v /home/yellowtent/platformdata/mysql:/var/lib/mysql                 --label isCloudronManaged=true                 --cap-add SYS_NICE                 --read-only -v /tmp -v /run registry.docker.com/cloudron/mysql:3.5.3@sha256:615057dbfeb5d979ab17c3e4cef240f11ae63b95feadbaf95a30e730472980aa
                                2026-04-24T17:54:34.909Z services: Waiting for mysql
                                2026-04-24T17:54:34.914Z services: Attempt 1 failed. Will retry: Network error waiting for mysql: connect ECONNREFUSED 172.18.30.1:3000
                                2026-04-24T17:54:50.061Z services: startPostgresql: stopping and deleting previous postgresql container
                                2026-04-24T17:54:50.061Z docker: stopContainer: stopping container postgresql
                                2026-04-24T17:54:50.063Z docker: deleteContainer: deleting postgresql
                                2026-04-24T17:54:50.063Z services: startPostgresql: starting postgresql container
                                2026-04-24T17:54:50.063Z shell: services: /bin/bash -c docker run --restart=unless-stopped -d --name=postgresql                 --hostname postgresql                 --net cloudron                 --net-alias postgresql                 --log-driver syslog                 --log-opt syslog-address=unix:///home/yellowtent/platformdata/logs/syslog.sock                 --log-opt syslog-format=rfc5424                 --log-opt tag=postgresql                 --ip 172.18.30.2                 --shm-size=128M                 -e CLOUDRON_POSTGRESQL_ROOT_PASSWORD=xxx                 -e CLOUDRON_POSTGRESQL_TOKEN=xxx                 -v /home/yellowtent/platformdata/postgresql:/var/lib/postgresql                 --label isCloudronManaged=true                 --read-only -v /tmp -v /run registry.docker.com/cloudron/postgresql:6.3.1@sha256:95a2e1e298fb88740c6d1d200657dcce993346ba500870e8e885a1aee519dc20
                                2026-04-24T17:54:50.455Z services: Waiting for postgresql
                                2026-04-24T17:54:50.458Z services: Attempt 1 failed. Will retry: Network error waiting for postgresql: connect ECONNREFUSED 172.18.30.2:3000
                                2026-04-24T17:55:05.613Z services: startMongodb: stopping and deleting previous mongodb container
                                2026-04-24T17:55:05.613Z docker: stopContainer: stopping container mongodb
                                2026-04-24T17:55:05.615Z docker: deleteContainer: deleting mongodb
                                2026-04-24T17:55:05.616Z shell: services: grep -q avx /proc/cpuinfo
                                2026-04-24T17:55:05.619Z services: startMongodb: starting mongodb container
                                2026-04-24T17:55:05.619Z shell: services: /bin/bash -c docker run --restart=unless-stopped -d --name=mongodb                 --hostname mongodb                 --net cloudron                 --net-alias mongodb                 --log-driver syslog                 --log-opt syslog-address=unix:///home/yellowtent/platformdata/logs/syslog.sock                 --log-opt syslog-format=rfc5424                 --log-opt tag=mongodb                 --ip 172.18.30.3                 -e CLOUDRON_MONGODB_ROOT_PASSWORD=xxx                 -e CLOUDRON_MONGODB_TOKEN=xxx                 -v /home/yellowtent/platformdata/mongodb:/var/lib/mongodb                 --label isCloudronManaged=true                 --read-only -v /tmp -v /run registry.docker.com/cloudron/mongodb:6.3.0@sha256:8757111970a99fb9a9880f02a2a77fe9cb7368745fcff555bf528471ecb50ec3
                                2026-04-24T17:55:05.995Z services: Waiting for mongodb
                                2026-04-24T17:55:05.998Z services: Attempt 1 failed. Will retry: Network error waiting for mongodb: connect ECONNREFUSED 172.18.30.3:3000
                                2026-04-24T17:55:21.502Z services: Attempt 2 failed. Will retry: Error waiting for mongodb. Status code: 200 message: Not ready yet: uptime=3
                                2026-04-24T17:55:36.979Z services: startRedis: stopping and deleting previous redis container redis-1502ffae-d38e-4b8a-a75d-c833b1355a4d
                                2026-04-24T17:55:36.979Z docker: stopContainer: stopping container redis-1502ffae-d38e-4b8a-a75d-c833b1355a4d
                                2026-04-24T17:55:36.981Z docker: deleteContainer: deleting redis-1502ffae-d38e-4b8a-a75d-c833b1355a4d
                                2026-04-24T17:55:36.982Z services: startRedis: starting redis container redis-1502ffae-d38e-4b8a-a75d-c833b1355a4d
                                2026-04-24T17:55:36.988Z shell: services: /bin/bash -c docker create --restart=unless-stopped --name=redis-1502ffae-d38e-4b8a-a75d-c833b1355a4d                 --hostname redis-1502ffae-d38e-4b8a-a75d-c833b1355a4d                 --label=location=data2.stirchley.coop                 --net cloudron                 --net-alias redis-1502ffae-d38e-4b8a-a75d-c833b1355a4d                 --log-driver syslog                 --log-opt syslog-address=unix:///home/yellowtent/platformdata/logs/syslog.sock                 --log-opt syslog-format=rfc5424                 --log-opt tag=redis-1502ffae-d38e-4b8a-a75d-c833b1355a4d                 -m 268435456                 --memory-swap -1                 -e CLOUDRON_REDIS_PASSWORD=xxx                 -e CLOUDRON_REDIS_TOKEN=xxx                 -v /home/yellowtent/platformdata/redis/1502ffae-d38e-4b8a-a75d-c833b1355a4d:/var/lib/redis                 --label isCloudronManaged=true                 --read-only -v /tmp -v /run registry.docker.com/cloudron/redis:3.8.0@sha256:54a12252edbc326fd4d10e8c00836cc68776f3463f22366279f7cd8d4dafd20f
                                2026-04-24T17:55:37.211Z services: startRedis: stopping and deleting previous redis container redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73
                                2026-04-24T17:55:37.211Z docker: stopContainer: stopping container redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73
                                2026-04-24T17:55:37.213Z docker: deleteContainer: deleting redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73
                                2026-04-24T17:55:37.213Z services: startRedis: starting redis container redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73
                                2026-04-24T17:55:37.217Z shell: services: /bin/bash -c docker run -d --restart=unless-stopped --name=redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73                 --hostname redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73                 --label=location=docs2.stirchley.coop                 --net cloudron                 --net-alias redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73                 --log-driver syslog                 --log-opt syslog-address=unix:///home/yellowtent/platformdata/logs/syslog.sock                 --log-opt syslog-format=rfc5424                 --log-opt tag=redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73                 -m 268435456                 --memory-swap -1                 -e CLOUDRON_REDIS_PASSWORD=                 -e CLOUDRON_REDIS_TOKEN=xxx                 -v /home/yellowtent/platformdata/redis/1aa55af7-48e4-4f6f-816f-46185ef4cf73:/var/lib/redis                 --label isCloudronManaged=true                 --read-only -v /tmp -v /run registry.docker.com/cloudron/redis:3.8.0@sha256:54a12252edbc326fd4d10e8c00836cc68776f3463f22366279f7cd8d4dafd20f
                                2026-04-24T17:55:37.465Z services: Waiting for redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73
                                2026-04-24T17:55:37.468Z services: Attempt 1 failed. Will retry: Network error waiting for redis-1aa55af7-48e4-4f6f-816f-46185ef4cf73: connect ECONNREFUSED 172.18.0.2:3000
                                
                                
                                1 Reply Last reply
                                0
                                • jamesJ Online
                                  jamesJ Online
                                  james
                                  Staff
                                  wrote last edited by james
                                  #16

                                  Apologies, I did a typo.

                                  systemctl restart box.service
                                  

                                  But I see you fixed the typo.

                                  1 Reply Last reply
                                  0
                                  • S Online
                                    S Online
                                    Stirchley.coop
                                    wrote last edited by
                                    #17

                                    That's fine I spotted and corrected it

                                    1 Reply Last reply
                                    0
                                    • jamesJ Online
                                      jamesJ Online
                                      james
                                      Staff
                                      wrote last edited by
                                      #18

                                      The log looks correct so far.
                                      It is recreating the containers due to the update and this can take some time.
                                      Can you observe this for maybe 5 minutes and report back?

                                      1 Reply Last reply
                                      0
                                      • S Online
                                        S Online
                                        Stirchley.coop
                                        wrote last edited by
                                        #19

                                        Sure will do

                                        1 Reply Last reply
                                        1
                                        • S Online
                                          S Online
                                          Stirchley.coop
                                          wrote last edited by
                                          #20

                                          It seems to have settled into this state:

                                          2026-04-24T18:01:40.032Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:01:50.030Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:02:00.043Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:02:10.027Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:02:20.032Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:02:30.028Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:02:40.034Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:02:50.028Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:03:00.040Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:03:10.027Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:03:20.029Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:03:30.027Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:03:40.030Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:03:50.031Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:04:00.037Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:04:10.029Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:04:20.035Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:04:30.029Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          2026-04-24T18:04:40.037Z apphealthmonitor: app health: 8 running / 5 stopped / 5 unresponsive
                                          

                                          There are 5 apps in Error state with the same mount error above. I assume they correspond to the 5 unresponsive apps in the logs. There are also 5 other stopped apps

                                          1 Reply Last reply
                                          1

                                          Hello! It looks like you're interested in this conversation, but you don't have an account yet.

                                          Getting fed up of having to scroll through the same posts each visit? When you register for an account, you'll always come back to exactly where you were before, and choose to be notified of new replies (either via email, or push notification). You'll also be able to save bookmarks and upvote posts to show your appreciation to other community members.

                                          With your input, this post could be even better 💗

                                          Register Login
                                          Reply
                                          • Reply as topic
                                          Log in to reply
                                          • Oldest to Newest
                                          • Newest to Oldest
                                          • Most Votes


                                          • Login

                                          • Don't have an account? Register

                                          • Login or register to search.
                                          • First post
                                            Last post
                                          0
                                          • Categories
                                          • Recent
                                          • Tags
                                          • Popular
                                          • Bookmarks
                                          • Search