Cloudron makes it easy to run web apps like WordPress, Nextcloud, GitLab on your server. Find out more or install now.


Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • Bookmarks
  • Search
Skins
  • Light
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
Brand Logo

Cloudron Forum

Apps | Demo | Docs | Install
  1. Cloudron Forum
  2. Support
  3. High memory usage spikes

High memory usage spikes

Scheduled Pinned Locked Moved Support
15 Posts 4 Posters 2.5k Views 4 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • robiR robi

    @subven nothing unusual on my end..

    syslogs show lots of packets dropped, so maybe a type of DDoS.

    subvenS Offline
    subvenS Offline
    subven
    wrote on last edited by
    #6

    @robi could be but what about the intervals? Very unusual for a DDoS....

    robiR 1 Reply Last reply
    0
    • subvenS subven

      @robi could be but what about the intervals? Very unusual for a DDoS....

      robiR Offline
      robiR Offline
      robi
      wrote on last edited by robi
      #7

      @subven agreed, there is no way to easily move from one app to another while looking at each apps graphs to compare network traffic.

      And the system info doesn't have a network graph at all.

      It happened again this morning at 9:31:07 and 10:16:07 with CPU spikes too:
      67b1a9fa-3557-4eed-b3e8-fcc53a47985d-image.png

      syslog has tons of messages from this IP blasting everyone:

      Jan 14 17:31:23 em kernel: [35147503.981778] Packet dropped: IN=enp3s0 OUT= MAC=
      ff:ff:ff:ff:ff:ff:52:54:a9:60:b1:6f:08:00 SRC=172.93.48.214 DST=255.255.255.255
      LEN=62 TOS=0x00 PREC=0x00 TTL=64 ID=34492 PROTO=UDP SPT=52322 DPT=8888 LEN=42
      

      Conscious tech

      1 Reply Last reply
      0
      • robiR Offline
        robiR Offline
        robi
        wrote on last edited by robi
        #8

        Started monitoring via top, 'm' to see when it spikes what is happening..
        this began taking 100% CPU:

        NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
        0 0 0 0 R 100.0 0.0 1:36.15 kworker/7:1+events_freezable
        0 0 0 0 S 55.8 0.0 9128:30 kswapd0

        g# free -h
                      total        used        free      shared  buff/cache   available
        Mem:           47Gi        44Gi       426Mi       228Mi       2.1Gi       1.7Gi
        Swap:         4.0Gi       1.4Gi       2.6Gi
        
        top - 22:11:56 up 406 days, 23:57,  3 users,  load average: 3.38, 2.69, 1.51
        Tasks: 477 total,   4 running, 473 sleeping,   0 stopped,   0 zombie
        %Cpu(s):  1.3 us, 14.1 sy,  0.0 ni, 84.4 id,  0.1 wa,  0.0 hi,  0.1 si,  0.0 st
        MiB Mem : 96.4/48195.9  [|||||||||||||||||||||||||||||||||||||||||||||||||||  ]
        MiB Swap: 35.0/4096.0   [|||||||||||||||||||                                  ]
        
            PID USER      PR  NI    VIRT    RES    SHR S  %CPU  %MEM     TIME+ COMMAND
        2357779 root      20   0       0      0      0 R 100.0   0.0   5:17.08 kworker+
            157 root      20   0       0      0      0 R  59.5   0.0   9129:10 kswapd0
         2449326 uuidd     20   0  393392  47296   5784 R  50.8   0.1 126:33.66 uwsgi
        
        

        What are these uwsgi processes?
        Seems like a python related memory leak needing atweak to the uswgi.ini?
        https://levelup.gitconnected.com/the-experience-of-solving-memory-leak-of-django-uwsgi-nginx-aws-cdb998244cfb

        b7ed071f-5a40-4637-8c2d-52f570e0f39d-image.png

        Conscious tech

        jdaviescoatesJ 1 Reply Last reply
        0
        • robiR robi

          Started monitoring via top, 'm' to see when it spikes what is happening..
          this began taking 100% CPU:

          NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
          0 0 0 0 R 100.0 0.0 1:36.15 kworker/7:1+events_freezable
          0 0 0 0 S 55.8 0.0 9128:30 kswapd0

          g# free -h
                        total        used        free      shared  buff/cache   available
          Mem:           47Gi        44Gi       426Mi       228Mi       2.1Gi       1.7Gi
          Swap:         4.0Gi       1.4Gi       2.6Gi
          
          top - 22:11:56 up 406 days, 23:57,  3 users,  load average: 3.38, 2.69, 1.51
          Tasks: 477 total,   4 running, 473 sleeping,   0 stopped,   0 zombie
          %Cpu(s):  1.3 us, 14.1 sy,  0.0 ni, 84.4 id,  0.1 wa,  0.0 hi,  0.1 si,  0.0 st
          MiB Mem : 96.4/48195.9  [|||||||||||||||||||||||||||||||||||||||||||||||||||  ]
          MiB Swap: 35.0/4096.0   [|||||||||||||||||||                                  ]
          
              PID USER      PR  NI    VIRT    RES    SHR S  %CPU  %MEM     TIME+ COMMAND
          2357779 root      20   0       0      0      0 R 100.0   0.0   5:17.08 kworker+
              157 root      20   0       0      0      0 R  59.5   0.0   9129:10 kswapd0
           2449326 uuidd     20   0  393392  47296   5784 R  50.8   0.1 126:33.66 uwsgi
          
          

          What are these uwsgi processes?
          Seems like a python related memory leak needing atweak to the uswgi.ini?
          https://levelup.gitconnected.com/the-experience-of-solving-memory-leak-of-django-uwsgi-nginx-aws-cdb998244cfb

          b7ed071f-5a40-4637-8c2d-52f570e0f39d-image.png

          jdaviescoatesJ Offline
          jdaviescoatesJ Offline
          jdaviescoates
          wrote on last edited by jdaviescoates
          #9

          @robi said in High memory usage spikes:

          kworker+

          Guess you've now seen these?

          https://www.linuxquestions.org/questions/linux-software-2/high-cpu-usage-by-kworker-4175563563/

          https://askubuntu.com/questions/33640/kworker-what-is-it-and-why-is-it-hogging-so-much-cpu

          Oh, and look, this person on SSDNodes also has a very similar issue:

          https://lowendtalk.com/discussion/169951/kworker-4-1-events-freezable-runs-every-hour-to-clean-buffer

          I'm gonna take this as yet another reason to not touch them with a barge pole! 😛

          I use Cloudron with Gandi & Hetzner

          robiR 1 Reply Last reply
          0
          • jdaviescoatesJ jdaviescoates

            @robi said in High memory usage spikes:

            kworker+

            Guess you've now seen these?

            https://www.linuxquestions.org/questions/linux-software-2/high-cpu-usage-by-kworker-4175563563/

            https://askubuntu.com/questions/33640/kworker-what-is-it-and-why-is-it-hogging-so-much-cpu

            Oh, and look, this person on SSDNodes also has a very similar issue:

            https://lowendtalk.com/discussion/169951/kworker-4-1-events-freezable-runs-every-hour-to-clean-buffer

            I'm gonna take this as yet another reason to not touch them with a barge pole! 😛

            robiR Offline
            robiR Offline
            robi
            wrote on last edited by
            #10

            @jdaviescoates that doesn't seem related..

            # cat /proc/vmstat | grep -E "dirty|writeback"
            nr_dirty 1166
            nr_writeback 0
            nr_writeback_temp 0
            nr_dirty_threshold 2192736
            nr_dirty_background_threshold 1095029
            

            I'll set echo 10000 > /proc/sys/vm/dirty_writeback_centisecs and see what happens.

            Conscious tech

            jdaviescoatesJ 1 Reply Last reply
            1
            • robiR robi

              @jdaviescoates that doesn't seem related..

              # cat /proc/vmstat | grep -E "dirty|writeback"
              nr_dirty 1166
              nr_writeback 0
              nr_writeback_temp 0
              nr_dirty_threshold 2192736
              nr_dirty_background_threshold 1095029
              

              I'll set echo 10000 > /proc/sys/vm/dirty_writeback_centisecs and see what happens.

              jdaviescoatesJ Offline
              jdaviescoatesJ Offline
              jdaviescoates
              wrote on last edited by jdaviescoates
              #11

              @robi in case you missed, I added a few other links, including this one more similar with someone also on SSDNodes (I think that's who you're using too, right?)

              https://lowendtalk.com/discussion/169951/kworker-4-1-events-freezable-runs-every-hour-to-clean-buffer

              I use Cloudron with Gandi & Hetzner

              robiR 1 Reply Last reply
              1
              • jdaviescoatesJ jdaviescoates

                @robi in case you missed, I added a few other links, including this one more similar with someone also on SSDNodes (I think that's who you're using too, right?)

                https://lowendtalk.com/discussion/169951/kworker-4-1-events-freezable-runs-every-hour-to-clean-buffer

                robiR Offline
                robiR Offline
                robi
                wrote on last edited by
                #12

                @jdaviescoates Yep, nice find.. it might be something they're doing. Will send them a note.

                Conscious tech

                1 Reply Last reply
                1
                • robiR Offline
                  robiR Offline
                  robi
                  wrote on last edited by robi
                  #13

                  Haven't been able to find where this lives:
                  2451531 ? Sl 127:28 5 0 393392 47284 0.0 uwsgi --master --workers 2 --buffer-size 16384 --no-orphans --ini /etc/uwsgi/apps-enabled/graphite-uwsgi.ini

                  and since when do we have such lengthy PIDs?

                  Conscious tech

                  girishG 1 Reply Last reply
                  0
                  • robiR robi

                    Haven't been able to find where this lives:
                    2451531 ? Sl 127:28 5 0 393392 47284 0.0 uwsgi --master --workers 2 --buffer-size 16384 --no-orphans --ini /etc/uwsgi/apps-enabled/graphite-uwsgi.ini

                    and since when do we have such lengthy PIDs?

                    girishG Offline
                    girishG Offline
                    girish
                    Staff
                    wrote on last edited by
                    #14

                    @robi that looks like the graphite container . Does that service have enough memory (under Services) ? You can also nuke /home/yellowtent/platformdata/graphite/whisper/collectd/localhost/* entirely and restart graphite. Maybe some corrupt whisper db file is causing issues.

                    robiR 1 Reply Last reply
                    1
                    • girishG girish

                      @robi that looks like the graphite container . Does that service have enough memory (under Services) ? You can also nuke /home/yellowtent/platformdata/graphite/whisper/collectd/localhost/* entirely and restart graphite. Maybe some corrupt whisper db file is causing issues.

                      robiR Offline
                      robiR Offline
                      robi
                      wrote on last edited by
                      #15

                      @girish Yes, plenty of memory, that is not the cause, but gets affected.

                      Graphite is not eating 40+GB of memory.

                      Conscious tech

                      1 Reply Last reply
                      0
                      Reply
                      • Reply as topic
                      Log in to reply
                      • Oldest to Newest
                      • Newest to Oldest
                      • Most Votes


                      • Login

                      • Don't have an account? Register

                      • Login or register to search.
                      • First post
                        Last post
                      0
                      • Categories
                      • Recent
                      • Tags
                      • Popular
                      • Bookmarks
                      • Search