SCIENTIFIC-LINUX-USERS Archives

September 2011

SCIENTIFIC-LINUX-USERS@LISTSERV.FNAL.GOV

Options: Use Monospaced Font
Show Text Part by Default
Show All Mail Headers

Message: [<< First] [< Prev] [Next >] [Last >>]
Topic: [<< First] [< Prev] [Next >] [Last >>]
Author: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
Subject:
From:
Stephan Wiesand <[log in to unmask]>
Reply To:
Stephan Wiesand <[log in to unmask]>
Date:
Mon, 5 Sep 2011 17:29:19 +0200
Content-Type:
text/plain
Parts/Attachments:
text/plain (122 lines)
Hi Stijn,

On Sep 5, 2011, at 17:23, Stijn De Weirdt wrote:

> hi stephan
> 
> the first lines of slabtop show
> 
> Active / Total Objects (% used)    : 238382 / 243482 (97.9%)
> Active / Total Slabs (% used)      : 4987 / 4987 (100.0%)
> Active / Total Caches (% used)     : 64 / 80 (80.0%)
> Active / Total Size (% used)       : 68080.12K / 69386.02K (98.1%)
> Minimum / Average / Maximum Object : 0.01K / 0.28K / 8.00K
> 
> is this similar?

not really.

% slabtop -s c --once |head -10
 Active / Total Objects (% used)    : 185625 / 205302 (90.4%)
 Active / Total Slabs (% used)      : 16756 / 16757 (100.0%)
 Active / Total Caches (% used)     : 101 / 182 (55.5%)
 Active / Total Size (% used)       : 856865.31K / 859451.65K (99.7%)
 Minimum / Average / Maximum Object : 0.02K / 4.19K / 4096.00K

  OBJS ACTIVE  USE OBJ SIZE  SLABS OBJ/SLAB CACHE SIZE NAME                   
   384    384 100% 2048.00K    384        1    786432K size-2097152           
 26276  26250  99%    1.00K   6569        4     26276K ext4_inode_cache       
   181    181 100%   32.12K    181        1     11584K kmem_cache             

> (btw size-2097152 sounds like one of the default name used by hugectl (or hugeadm) from the hugetlbfs tools). is that mounted in your case? and are there any hugepages reserved? )

Not that I'd know of. But wasn't there a new feature called "transparent hugepage support" in 6.1?

Cheers,
	Stephan
> 
> 
> stijn
> 
> On 09/05/2011 05:10 PM, Stephan Wiesand wrote:
>> Hi Stijn,
>> 
>> On Sep 5, 2011, at 16:24, Stijn De Weirdt wrote:
>> 
>>> hi all,
>>> 
>>> we are having an "issue" with some SL61 nodes. after a reboot, free reports 1.4GB of memory in use, of which 24+163=187MB buffers+cache.
>>> 
>>> i'm unable to identify what is holding the memory, and i'd like to know if others see this too and how i could proceed to find the culprit.
>> 
>> yes, we see this as well. On a 48 GB system without users or special processes:
>> 
>> # free -m
>>              total       used       free     shared    buffers     cached
>> Mem:         48388       1374      47013          0         30        186
>> -/+ buffers/cache:       1157      47231
>> 
>> In /proc/meminfo, I find that the difference to what I'd consider reasonable (and see on a 48GB SL5 system) is due to slabs.
>> 
>> A "slabtop -s c" reveals that it's a "size-2097152" pool accounting for this. Do you see this as well?
>> 
>> Cheers,
>> 	Stephan
>> 
>>> 
>>> (it is a 32core/64GB machine; kernel commandline has crashkernel=128M@16M (but no difference then eg crashkernel=auto and kdump is off))
>>> 
>>> many thanks,
>>> 
>>> 
>>> stijn
>>> 
>>> free
>>> # free -m
>>>             total       used       free     shared    buffers     cached
>>> Mem:         64554       1604      62949          0         24        166
>>> -/+ buffers/cache:       1413      63140
>>> Swap:        16394          0      16394
>>> 
>>> 
>>> mem sorted top
>>> 
>>> top - 16:13:52 up 13 min,  1 user,  load average: 0.00, 0.01, 0.01
>>> Tasks: 694 total,   1 running, 693 sleeping,   0 stopped,   0 zombie
>>> Cpu(s):  0.0%us,  0.0%sy,  0.0%ni,100.0%id,  0.0%wa,  0.0%hi,  0.0%si, 0.0%st
>>> Mem:  66103768k total,  1643336k used, 64460432k free,    25164k buffers
>>> Swap: 16787916k total,        0k used, 16787916k free,   170552k cached
>>> 
>>>  PID USER      PR  NI  VIRT  RES  SHR S %CPU %MEM    TIME+  COMMAND
>>> 2788 root      20   0 37988  25m 2876 S  0.0  0.0   0:00.06 pbs_mom
>>> 2653 root      20   0  159m  12m 1472 S  0.0  0.0   0:00.19 ncm-cdispd
>>> 2643 root      20   0  138m 5604  840 S  0.0  0.0   0:00.00 cdp-listend
>>> 3276 root      20   0  120m 4156 3232 S  0.0  0.0   0:00.07 sshd
>>> 2620 root      20   0  745m 3788 1764 S  0.0  0.0   0:00.12 automount
>>> 3102 nslcd     20   0  427m 2936  488 S  0.0  0.0   0:00.00 nslcd
>>> 3301 root      20   0  103m 1688 1336 S  0.0  0.0   0:00.05 bash
>>> 3623 root      20   0 13528 1604  844 R  0.3  0.0   0:00.14 top
>>>    1 root      20   0 21416 1544 1240 S  0.0  0.0   0:06.23 init
>>> 2482 root      20   0  194m 1484 1108 S  0.0  0.0   0:00.14 qlgc_dsc
>>> 2325 root      20   0  242m 1412  928 S  0.0  0.0   0:00.04 rsyslogd
>>> 2459 rpcuser   20   0 23112 1168  884 S  0.0  0.0   0:00.00 rpc.statd
>>> 2606 root      18  -2 10956 1144  412 S  0.0  0.0   0:00.03 udevd
>>> 3164 nscd      20   0  583m 1132  788 S  0.0  0.0   0:00.02 nscd
>>> 2697 root      20   0 62040 1064  464 S  0.0  0.0   0:00.00 sshd
>>>  943 root      16  -4 10960 1052  316 S  0.0  0.0   0:00.12 udevd
>>> 2607 root      18  -2 10956 1052  320 S  0.0  0.0   0:00.00 udevd
>>> 2723 root      20   0  112m 1012  380 S  0.0  0.0   0:00.00 crond
>>> 2707 root      20   0 22488  992  752 S  0.0  0.0   0:00.03 xinetd
>>> 2439 rpc       20   0 18940  908  672 S  0.0  0.0   0:00.04 rpcbind
>>> 2568 dbus      20   0 23448  876  604 S  0.0  0.0   0:00.01 dbus-daemon
>>> 2972 nagios    20   0 37096  796  452 S  0.0  0.0   0:00.00 nrpe
>> 
> 
> 

-- 
Stephan Wiesand
DESY -DV-
Platanenallee 6
15738 Zeuthen, Germany

ATOM RSS1 RSS2