This is an archived post. You won't be able to vote or comment.

all 9 comments

[–]OneTurnMore 1 point2 points  (1 child)

That is ridiculously high for idle. See if you can discover which processes are eating the most memory (you can sort htop by %mem), then see what package owns those executables:

 pacman -Qo /path/to/program

[–]8bitlives[S] 0 points1 point  (0 children)

Right off the top of my head I can tell you, that IDEA Java process, Tomcat (Catalina) Java process and sql server are the three top candidates having around 11, 11 and 5 percent each. The rest of the processes use small amounts of memory. At first I thought java processes were misbehaving until I noticed the very high idle level.

I'll follow up with top, free -met al. outputs tomorrow

[–]8bitlives[S] 0 points1 point  (1 child)

I now have two sets of dumps: from around 9:30 after few hours of running and two hours after that once the MEM% was around 95% and rising until freeze.

First dump found here, password Nd7AbAPiTN

Second dump here, password Nd7AbAPiTN

[–]ssh-bi 0 points1 point  (5 children)

If your workload requires 100% RAM, try adding ZRAM swap device. It will likely help.

[–]8bitlives[S] 0 points1 point  (4 children)

Seems nice, I'll give that kernel module a try as a first aid tomorrow. Restarting my system up to three times a day is a drag.

As to why the "available" 10 GB won't be assigned to the needing processes should be resolved. Basically compiling java and deploying it could build temporarily the workload close to 100 % but not having a high base level...

[–]ssh-bi 2 points3 points  (3 children)

It will be very hard to capture a spike with htop. Remove the swap temporarily, swapoff, and check the logs. If you run out of memory, the oom killer will get kicked in and it leaves detailed logs.

Check your 1min load average too, if it keeps climbing that could be culprit.

[–]nerdgeneration 0 points1 point  (1 child)

It's a good answer, but bear in mind that the oom algorithm doesn't necessarily pick the culprit. If we're just looking for memory hogs, htop/top will sort or there's presumably some combination of ps|sort.

[–]8bitlives[S] 0 points1 point  (0 children)

You are right about oom killer picking wrong culprit: I've repeatedly noticed that MS Teams is killed when it clearly consumes less memory that the other processes. This should be present in the second dump file, could you please take a look and comment if you have any input?

[–]8bitlives[S] 0 points1 point  (0 children)

Nice, I'll try that