SCIENTIFIC-LINUX-USERS Archives

September 2011

SCIENTIFIC-LINUX-USERS@LISTSERV.FNAL.GOV

Options: Use Monospaced Font
Show Text Part by Default
Show All Mail Headers

Message: [<< First] [< Prev] [Next >] [Last >>]
Topic: [<< First] [< Prev] [Next >] [Last >>]
Author: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
Subject:
From:
Stephan Wiesand <[log in to unmask]>
Reply To:
Stephan Wiesand <[log in to unmask]>
Date:
Mon, 5 Sep 2011 17:10:38 +0200
Content-Type:
text/plain
Parts/Attachments:
text/plain (78 lines)
Hi Stijn,

On Sep 5, 2011, at 16:24, Stijn De Weirdt wrote:

> hi all,
> 
> we are having an "issue" with some SL61 nodes. after a reboot, free reports 1.4GB of memory in use, of which 24+163=187MB buffers+cache.
> 
> i'm unable to identify what is holding the memory, and i'd like to know if others see this too and how i could proceed to find the culprit.

yes, we see this as well. On a 48 GB system without users or special processes:

# free -m
             total       used       free     shared    buffers     cached
Mem:         48388       1374      47013          0         30        186
-/+ buffers/cache:       1157      47231

In /proc/meminfo, I find that the difference to what I'd consider reasonable (and see on a 48GB SL5 system) is due to slabs.

A "slabtop -s c" reveals that it's a "size-2097152" pool accounting for this. Do you see this as well?

Cheers,
	Stephan

> 
> (it is a 32core/64GB machine; kernel commandline has crashkernel=128M@16M (but no difference then eg crashkernel=auto and kdump is off))
> 
> many thanks,
> 
> 
> stijn
> 
> free
> # free -m
>             total       used       free     shared    buffers     cached
> Mem:         64554       1604      62949          0         24        166
> -/+ buffers/cache:       1413      63140
> Swap:        16394          0      16394
> 
> 
> mem sorted top
> 
> top - 16:13:52 up 13 min,  1 user,  load average: 0.00, 0.01, 0.01
> Tasks: 694 total,   1 running, 693 sleeping,   0 stopped,   0 zombie
> Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si, 0.0%st
> Mem:  66103768k total,  1643336k used, 64460432k free,    25164k buffers
> Swap: 16787916k total,        0k used, 16787916k free,   170552k cached
> 
>  PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND 
> 2788 root      20   0 37988  25m 2876 S  0.0  0.0   0:00.06 pbs_mom 
> 2653 root      20   0  159m  12m 1472 S  0.0  0.0   0:00.19 ncm-cdispd 
> 2643 root      20   0  138m 5604  840 S  0.0  0.0   0:00.00 cdp-listend 
> 3276 root      20   0  120m 4156 3232 S  0.0  0.0   0:00.07 sshd 
> 2620 root      20   0  745m 3788 1764 S  0.0  0.0   0:00.12 automount 
> 3102 nslcd     20   0  427m 2936  488 S  0.0  0.0   0:00.00 nslcd 
> 3301 root      20   0  103m 1688 1336 S  0.0  0.0   0:00.05 bash 
> 3623 root      20   0 13528 1604  844 R  0.3  0.0   0:00.14 top 
>    1 root      20   0 21416 1544 1240 S  0.0  0.0   0:06.23 init 
> 2482 root      20   0  194m 1484 1108 S  0.0  0.0   0:00.14 qlgc_dsc 
> 2325 root      20   0  242m 1412  928 S  0.0  0.0   0:00.04 rsyslogd 
> 2459 rpcuser   20   0 23112 1168  884 S  0.0  0.0   0:00.00 rpc.statd 
> 2606 root      18  -2 10956 1144  412 S  0.0  0.0   0:00.03 udevd 
> 3164 nscd      20   0  583m 1132  788 S  0.0  0.0   0:00.02 nscd 
> 2697 root      20   0 62040 1064  464 S  0.0  0.0   0:00.00 sshd 
>  943 root      16  -4 10960 1052  316 S  0.0  0.0   0:00.12 udevd 
> 2607 root      18  -2 10956 1052  320 S  0.0  0.0   0:00.00 udevd 
> 2723 root      20   0  112m 1012  380 S  0.0  0.0   0:00.00 crond 
> 2707 root      20   0 22488  992  752 S  0.0  0.0   0:00.03 xinetd 
> 2439 rpc       20   0 18940  908  672 S  0.0  0.0   0:00.04 rpcbind 
> 2568 dbus      20   0 23448  876  604 S  0.0  0.0   0:00.01 dbus-daemon 
> 2972 nagios    20   0 37096  796  452 S  0.0  0.0   0:00.00 nrpe 

-- 
Stephan Wiesand
DESY -DV-
Platanenallee 6
15738 Zeuthen, Germany

ATOM RSS1 RSS2