For the rest: The issue is that there are a large number of files (> 400k) getting backed up and the syncer code is running out of memory since it loads all those filenames in memory. I will get this fixed in the next version. For now, disable automatic backups for nextcloud.
Generally doing any additional system configuration or removing/adding other ubuntu packages to the system is not supported, since we cannot test such variations for updates.
Cloudron already only installs signed packages. Enabling livepatch should be ok to do.
For all the other things happening through that ansible role, we would have to go through them one by one and test accordingly. We will not support running such hardening scripts automatically, there are too many of these out there. So if there are really good reasons to disable/configure system components for security we can investigate. Often security roles don't even apply to Cloudron if the corresponding components are not even used.
@relink Cloudron does not work without a domain name. Many features like DNS integration, certificate management, reverse proxy setup etc all require a domain name.
Maybe you can try a setup like this:
Pick an imaginary domain name like relink.home
Choose the noop (only for development) DNS provider.
In Advanced section, choose Self signed certificates
In your router/DNS, just add entries for *.relink.home and relink.home to the server's IP. Alternately, add those entires to the /etc/hosts of your PC. You should then be able to reach, my.relink.home and install apps.
Some users have reported issues with logging into SOGo even when using the Cloudron email after the update. This is most likely because you are using an older version of SOGo. The old SOGo app was deprecated ~6 months ago as part of the 2.3 release and is not supported. Please move to the new SOGo immediately.
Install the latest version of the SOGo app (from the appstore) in another location/subdomain. You should see your email when logging in with email now.
@CarbonBee Did I understand correctly that the DNS of cloudron A and Cloudron B are different above?
If so, the behavior is expected (not saying it is correct or even good). Currently, there is no easy way to easily 'migrate' domains - it is only easy to migrate a Cloudron from one server to another provided that all the DNS remains the same.
Can you explain your use case a bit more so we can try to come up with a solution together? I guess you are trying to test your backups? How can we make this work when Cloudron has multiple domains? Once you restore to another server, Cloudron will do the DNS setup of apps automatically and it will end up re-configuring the DNS of the apps to point to this new server.
@robw Sorry for the delayed response, we are just coming back from vacation and catching up on support tickets.
If I understand correctly, the Cloudron server has a different outbound IP than the one it detects. We have a custom endpoint (https://api.cloudron.io/api/v1/helper/public_ip) which helps us detect the IP of a server but I guess this detection goes wrong because of your setup.
To fix/workaround this: In the domain setup wizard, you can simply choose "no-op" as the DNS provider. With this provider, all DNS checks are disabled and as long as the domain somehow is able to resolve and reach your cloudron, it should all work. Another thing is that port 80 needs to be reachable as well for Let's Encrypt to work. If this is not possible, you can select 'Self Signed Certs' in the Advanced section of the domain UI.