SCIENTIFIC-LINUX-USERS Archives

September 2011

SCIENTIFIC-LINUX-USERS@LISTSERV.FNAL.GOV

Options: Use Monospaced Font
Show Text Part by Default
Show All Mail Headers

Message: [<< First] [< Prev] [Next >] [Last >>]
Topic: [<< First] [< Prev] [Next >] [Last >>]
Author: [<< First] [< Prev] [Next >] [Last >>]

Print Reply
Subject:
From:
Stijn De Weirdt <[log in to unmask]>
Reply To:
Stijn De Weirdt <[log in to unmask]>
Date:
Thu, 8 Sep 2011 15:58:36 +0200
Content-Type:
text/plain
Parts/Attachments:
text/plain (142 lines)
issue resolved (somewhat)


dmesg|Memory

shows 1GB reserved (of which 850MB memmap = 1.3% of total memory)
it also shows 650MB absent, apparently this is due to holes in the 
memory mapping, which is something hardware specific (it's an AMD Dell 
C6145, so it's amd specific or Dell/Bios specific)

anyway, thanks for the feedback

stijn

>> % slabtop -s c --once |head -10
>> Active / Total Objects (% used) : 185625 / 205302 (90.4%)
>> Active / Total Slabs (% used) : 16756 / 16757 (100.0%)
>> Active / Total Caches (% used) : 101 / 182 (55.5%)
>> Active / Total Size (% used) : 856865.31K / 859451.65K (99.7%)
>> Minimum / Average / Maximum Object : 0.02K / 4.19K / 4096.00K
>>
>> OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME
>> 384 384 100% 2048.00K 384 1 786432K size-2097152
>> 26276 26250 99% 1.00K 6569 4 26276K ext4_inode_cache
>> 181 181 100% 32.12K 181 1 11584K kmem_cache
> very different indeed.
>
>>
>>> (btw size-2097152 sounds like one of the default name used by hugectl
>>> (or hugeadm) from the hugetlbfs tools). is that mounted in your case?
>>> and are there any hugepages reserved? )
>>
>> Not that I'd know of. But wasn't there a new feature called
>> "transparent hugepage support" in 6.1?
> thp shows in /proc/meminfo as AnonHugePages
>
> ]# grep Huge /proc/meminfo
> AnonHugePages: 18432 kB
> HugePages_Total: 0
> HugePages_Free: 0
> HugePages_Rsvd: 0
> HugePages_Surp: 0
> Hugepagesize: 2048 kB
>
> you show check those there too to see where the come from
>
>
> stijn
>
>>
>> Cheers,
>> Stephan
>>>
>>>
>>> stijn
>>>
>>> On 09/05/2011 05:10 PM, Stephan Wiesand wrote:
>>>> Hi Stijn,
>>>>
>>>> On Sep 5, 2011, at 16:24, Stijn De Weirdt wrote:
>>>>
>>>>> hi all,
>>>>>
>>>>> we are having an "issue" with some SL61 nodes. after a reboot, free
>>>>> reports 1.4GB of memory in use, of which 24+163=187MB buffers+cache.
>>>>>
>>>>> i'm unable to identify what is holding the memory, and i'd like to
>>>>> know if others see this too and how i could proceed to find the
>>>>> culprit.
>>>>
>>>> yes, we see this as well. On a 48 GB system without users or special
>>>> processes:
>>>>
>>>> # free -m
>>>> total used free shared buffers cached
>>>> Mem: 48388 1374 47013 0 30 186
>>>> -/+ buffers/cache: 1157 47231
>>>>
>>>> In /proc/meminfo, I find that the difference to what I'd consider
>>>> reasonable (and see on a 48GB SL5 system) is due to slabs.
>>>>
>>>> A "slabtop -s c" reveals that it's a "size-2097152" pool accounting
>>>> for this. Do you see this as well?
>>>>
>>>> Cheers,
>>>> Stephan
>>>>
>>>>>
>>>>> (it is a 32core/64GB machine; kernel commandline has
>>>>> crashkernel=128M@16M (but no difference then eg crashkernel=auto
>>>>> and kdump is off))
>>>>>
>>>>> many thanks,
>>>>>
>>>>>
>>>>> stijn
>>>>>
>>>>> free
>>>>> # free -m
>>>>> total used free shared buffers cached
>>>>> Mem: 64554 1604 62949 0 24 166
>>>>> -/+ buffers/cache: 1413 63140
>>>>> Swap: 16394 0 16394
>>>>>
>>>>>
>>>>> mem sorted top
>>>>>
>>>>> top - 16:13:52 up 13 min, 1 user, load average: 0.00, 0.01, 0.01
>>>>> Tasks: 694 total, 1 running, 693 sleeping, 0 stopped, 0 zombie
>>>>> Cpu(s): 0.0%us, 0.0%sy, 0.0%ni,100.0%id, 0.0%wa, 0.0%hi, 0.0%si,
>>>>> 0.0%st
>>>>> Mem: 66103768k total, 1643336k used, 64460432k free, 25164k buffers
>>>>> Swap: 16787916k total, 0k used, 16787916k free, 170552k cached
>>>>>
>>>>> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND
>>>>> 2788 root 20 0 37988 25m 2876 S 0.0 0.0 0:00.06 pbs_mom
>>>>> 2653 root 20 0 159m 12m 1472 S 0.0 0.0 0:00.19 ncm-cdispd
>>>>> 2643 root 20 0 138m 5604 840 S 0.0 0.0 0:00.00 cdp-listend
>>>>> 3276 root 20 0 120m 4156 3232 S 0.0 0.0 0:00.07 sshd
>>>>> 2620 root 20 0 745m 3788 1764 S 0.0 0.0 0:00.12 automount
>>>>> 3102 nslcd 20 0 427m 2936 488 S 0.0 0.0 0:00.00 nslcd
>>>>> 3301 root 20 0 103m 1688 1336 S 0.0 0.0 0:00.05 bash
>>>>> 3623 root 20 0 13528 1604 844 R 0.3 0.0 0:00.14 top
>>>>> 1 root 20 0 21416 1544 1240 S 0.0 0.0 0:06.23 init
>>>>> 2482 root 20 0 194m 1484 1108 S 0.0 0.0 0:00.14 qlgc_dsc
>>>>> 2325 root 20 0 242m 1412 928 S 0.0 0.0 0:00.04 rsyslogd
>>>>> 2459 rpcuser 20 0 23112 1168 884 S 0.0 0.0 0:00.00 rpc.statd
>>>>> 2606 root 18 -2 10956 1144 412 S 0.0 0.0 0:00.03 udevd
>>>>> 3164 nscd 20 0 583m 1132 788 S 0.0 0.0 0:00.02 nscd
>>>>> 2697 root 20 0 62040 1064 464 S 0.0 0.0 0:00.00 sshd
>>>>> 943 root 16 -4 10960 1052 316 S 0.0 0.0 0:00.12 udevd
>>>>> 2607 root 18 -2 10956 1052 320 S 0.0 0.0 0:00.00 udevd
>>>>> 2723 root 20 0 112m 1012 380 S 0.0 0.0 0:00.00 crond
>>>>> 2707 root 20 0 22488 992 752 S 0.0 0.0 0:00.03 xinetd
>>>>> 2439 rpc 20 0 18940 908 672 S 0.0 0.0 0:00.04 rpcbind
>>>>> 2568 dbus 20 0 23448 876 604 S 0.0 0.0 0:00.01 dbus-daemon
>>>>> 2972 nagios 20 0 37096 796 452 S 0.0 0.0 0:00.00 nrpe
>>>>
>>>
>>>
>>

ATOM RSS1 RSS2