SCIENTIFIC-LINUX-USERS Archives

September 2011

SCIENTIFIC-LINUX-USERS@LISTSERV.FNAL.GOV

Options: Use Monospaced Font
Show Text Part by Default
Show All Mail Headers

Message: [<< First] [< Prev] [Next >] [Last >>]
Topic: [<< First] [< Prev] [Next >] [Last >>]
Author: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
Subject:
From:
Stijn De Weirdt <[log in to unmask]>
Reply To:
Stijn De Weirdt <[log in to unmask]>
Date:
Mon, 5 Sep 2011 17:23:18 +0200
Content-Type:
text/plain
Parts/Attachments:
text/plain (93 lines)
hi stephan

the first lines of slabtop show

Active / Total Objects (% used)    : 238382 / 243482 (97.9%)
  Active / Total Slabs (% used)      : 4987 / 4987 (100.0%)
  Active / Total Caches (% used)     : 64 / 80 (80.0%)
  Active / Total Size (% used)       : 68080.12K / 69386.02K (98.1%)
  Minimum / Average / Maximum Object : 0.01K / 0.28K / 8.00K

is this similar?

(btw size-2097152 sounds like one of the default name used by hugectl 
(or hugeadm) from the hugetlbfs tools). is that mounted in your case? 
and are there any hugepages reserved? )


stijn

On 09/05/2011 05:10 PM, Stephan Wiesand wrote:
> Hi Stijn,
>
> On Sep 5, 2011, at 16:24, Stijn De Weirdt wrote:
>
>> hi all,
>>
>> we are having an "issue" with some SL61 nodes. after a reboot, free reports 1.4GB of memory in use, of which 24+163=187MB buffers+cache.
>>
>> i'm unable to identify what is holding the memory, and i'd like to know if others see this too and how i could proceed to find the culprit.
>
> yes, we see this as well. On a 48 GB system without users or special processes:
>
> # free -m
>               total       used       free     shared    buffers     cached
> Mem:         48388       1374      47013          0         30        186
> -/+ buffers/cache:       1157      47231
>
> In /proc/meminfo, I find that the difference to what I'd consider reasonable (and see on a 48GB SL5 system) is due to slabs.
>
> A "slabtop -s c" reveals that it's a "size-2097152" pool accounting for this. Do you see this as well?
>
> Cheers,
> 	Stephan
>
>>
>> (it is a 32core/64GB machine; kernel commandline has crashkernel=128M@16M (but no difference then eg crashkernel=auto and kdump is off))
>>
>> many thanks,
>>
>>
>> stijn
>>
>> free
>> # free -m
>>              total       used       free     shared    buffers     cached
>> Mem:         64554       1604      62949          0         24        166
>> -/+ buffers/cache:       1413      63140
>> Swap:        16394          0      16394
>>
>>
>> mem sorted top
>>
>> top - 16:13:52 up 13 min,  1 user,  load average: 0.00, 0.01, 0.01
>> Tasks: 694 total,   1 running, 693 sleeping,   0 stopped,   0 zombie
>> Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si, 0.0%st
>> Mem:  66103768k total,  1643336k used, 64460432k free,    25164k buffers
>> Swap: 16787916k total,        0k used, 16787916k free,   170552k cached
>>
>>   PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
>> 2788 root      20   0 37988  25m 2876 S  0.0  0.0   0:00.06 pbs_mom
>> 2653 root      20   0  159m  12m 1472 S  0.0  0.0   0:00.19 ncm-cdispd
>> 2643 root      20   0  138m 5604  840 S  0.0  0.0   0:00.00 cdp-listend
>> 3276 root      20   0  120m 4156 3232 S  0.0  0.0   0:00.07 sshd
>> 2620 root      20   0  745m 3788 1764 S  0.0  0.0   0:00.12 automount
>> 3102 nslcd     20   0  427m 2936  488 S  0.0  0.0   0:00.00 nslcd
>> 3301 root      20   0  103m 1688 1336 S  0.0  0.0   0:00.05 bash
>> 3623 root      20   0 13528 1604  844 R  0.3  0.0   0:00.14 top
>>     1 root      20   0 21416 1544 1240 S  0.0  0.0   0:06.23 init
>> 2482 root      20   0  194m 1484 1108 S  0.0  0.0   0:00.14 qlgc_dsc
>> 2325 root      20   0  242m 1412  928 S  0.0  0.0   0:00.04 rsyslogd
>> 2459 rpcuser   20   0 23112 1168  884 S  0.0  0.0   0:00.00 rpc.statd
>> 2606 root      18  -2 10956 1144  412 S  0.0  0.0   0:00.03 udevd
>> 3164 nscd      20   0  583m 1132  788 S  0.0  0.0   0:00.02 nscd
>> 2697 root      20   0 62040 1064  464 S  0.0  0.0   0:00.00 sshd
>>   943 root      16  -4 10960 1052  316 S  0.0  0.0   0:00.12 udevd
>> 2607 root      18  -2 10956 1052  320 S  0.0  0.0   0:00.00 udevd
>> 2723 root      20   0  112m 1012  380 S  0.0  0.0   0:00.00 crond
>> 2707 root      20   0 22488  992  752 S  0.0  0.0   0:00.03 xinetd
>> 2439 rpc       20   0 18940  908  672 S  0.0  0.0   0:00.04 rpcbind
>> 2568 dbus      20   0 23448  876  604 S  0.0  0.0   0:00.01 dbus-daemon
>> 2972 nagios    20   0 37096  796  452 S  0.0  0.0   0:00.00 nrpe
>

ATOM RSS1 RSS2