[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]

Re: Out of Memory issue






Hello Eric,
                      Thanks for your detailed reply. I have looked at the
things that you suggested, unfortunately upgrading to 64bit is not an
option for us at the moment due to software compatibility. I have looked in
the /proc/sys/vm directory and i do not currently have a
lower_zone_protection file or an oom-kill file, also these cannot be added
manually, maybe this is because i am running 2.4.21.20 kernel. Out of the
options you gave me, the upgrade to the hugemem kernel seems the only one I
can try currently. I do not have access the redhat network currently but
managed to find a "kernel-hugemem-unsupported-2.4.21-4.EL.i686.rpm" on the
installation cd. This version is few versions older than the one i am
currently running but decided to install it anyway. The installation seemed
to go fine but i am not getting the option on boot to choose the kernel i
installed.

Is this because it is older than the one i already have? Or is there
something else i need to do to get grub to see it on boot? Also what is the
latest version of my current 2.4.21.?? kernel?

P.s Does anyone know where i can get the latest 2.4.21.?? kernel apart from
the redhat network website.

Any help on this would be much appreciated.

Regards

Andrew Bridgeman
|+----------------------------+-------------------------------------------|
||   Eric Sisler              |                                           |
||   <esisler westminster lib |           To:        Red Hat Linux        |
||   co.us>                   |   discussion list <redhat-list redhat com>|
||   Sent by:                 |           cc:                             |
||   redhat-list-bounces redha|           Subject:        Re: Out of      |
||   t.com                    |   Memory issue                            |
||                            |                                           |
||   17/01/2008 16:27         |                                           |
||   Please respond to General|                                           |
||   Red Hat Linux discussion |                                           |
||   list                     |                                           |
||                            |                                           |
|+----------------------------+-------------------------------------------|







Andrew Bridgeman corusgroup com wrote:

> I currently have a problem with our Redhat Machines running out of memory
> and automatically killing the main process that is running on the machine
> at the time. Below is my kernel version and the text from the messages
> file. I have checked the top command and the machine seems to have loads
of
> the 2 gig memory left. Could someone please let me know how to resolve
this
> issue.

This sounds like the infamous "OOM killer" problem.  From an e-mail I
posted around August of 2007:

Since this problem seems to popup on different lists, this message has
been cross-posted to the general Red Hat discussion list, the RHEL3
(Taroon) list and the RHEL4 (Nahant) list.  My apologies for not having
the time to post this summary sooner.

I would still be banging my head against this problem were it not for
the generous assistance of Tom Sightler <ttsig tuxyturvy com> and Brian
Long <brilong cisco com>.

In general, the out of memory killer (oom-killer) begins killing
processes, even on servers with large amounts (6Gb+) of RAM.  In many
cases people report plenty of "free" RAM and are perplexed as to why the
oom-killer is whacking processes.  Indications that this has happened
appear in /var/log/messages:
Out of Memory: Killed process [PID] [process name].

In my case I was upgrading various VMware servers from RHEL3 / VMware
GSX to RHEL4 / VMware Server.  One of the virtual machines on a server
with 16Gb of RAM kept getting whacked by the oom-killer.  Needless to
say, this was quite frustrating.

As it turns out, the problem was low memory exhaustion.  Quoting Tom:
"The kernel uses low memory to track allocations of all memory thus a
system with 16GB of memory will use significantly more low memory than a
system with 4GB, perhaps as much as 4 times.  This extra pressure
happens from the moment you turn the system on before you do anything at
all because the kernel structures have to be sized for the potential of
tracking allocations in four times as much memory."

You can check the status of low & high memory a couple of ways:

# egrep 'High|Low' /proc/meminfo
HighTotal:     5111780 kB
HighFree:         1172 kB
LowTotal:       795688 kB
LowFree:         16788 kB

# free -lm
total       used       free     shared    buffers    cached
Mem:          5769       5751         17          0          8      5267
Low:           777        760         16          0          0         0
High:         4991       4990          1          0          0         0
-/+ buffers/cache:        475       5293
Swap:         4773          0       4773

When low memory is exhausted, it doesn't matter how much high memory is
available, the oom-killer will begin whacking processes to keep the
server alive.

There are a couple of solutions to this problem:

If possible, upgrade to 64-bit Linux.  This is the best solution because
*all* memory becomes low memory.  If you run out of low memory in this
case, then you're *really* out of memory. ;-)

If limited to 32-bit Linux, the best solution is to run the hugemem
kernel.  This kernel splits low/high memory differently, and in most
cases should provide enough low memory to map high memory.  In most
cases this is an easy fix - simply install the hugemem kernel RPM &
reboot.

If running the 32-bit hugemem kernel isn't an option either, you can try
setting /proc/sys/vm/lower_zone_protection to a value of 250 or more.
This will cause the kernel to try to be more aggressive in defending the
low zone from allocating memory that could potentially be allocated in
the high memory zone.  As far as I know, this option isn't available
until the 2.6.x kernel. Some experimentation to find the best setting
for your environment will probably be necessary.  You can check & set
this value on the fly via:
# cat /proc/sys/vm/lower_zone_protection
# echo "250" > /proc/sys/vm/lower_zone_protection

To set this option on boot, add the following to /etc/sysctl.conf:
vm.lower_zone_protection = 250

As a last-ditch effort, you can disable the oom-killer.  This option can
cause the server to hang, so use it with extreme caution (and at your
own risk)!
Check status of oom-killer:
# cat /proc/sys/vm/oom-kill

Turn oom-killer off/on:
# echo "0" > /proc/sys/vm/oom-kill
# echo "1" > /proc/sys/vm/oom-kill

To make this change take effect at boot time, add the following
to /etc/sysctl.conf:
vm.oom-kill = 0

For processes that would have been killed, but weren't because the oom-
killer is disabled, you'll see the following message
in /var/log/messages:
"Would have oom-killed but /proc/sys/vm/oom-kill is disabled"

Sorry for being so long-winded.  I hope this helps others who have
struggled with this problem.

-Eric

--
Eric Sisler <esisler westminster lib co us>
Library Network Specialist
Westminster Public Library
Westminster, CO USA

Linux - Don't fear the Penguin.
Want to know what we use Linux for?
Visit http://wallace.westminster.lib.co.us/linux

--
redhat-list mailing list
unsubscribe mailto:redhat-list-request redhat com?subject=unsubscribe
https://www.redhat.com/mailman/listinfo/redhat-list


**********************************************************************
This transmission is confidential and must not be used or disclosed by
anyone other than the intended recipient. Neither Corus Group Limited nor
any of its subsidiaries can accept any responsibility for any use or
misuse of the transmission by anyone.

For address and company registration details of certain entities
within the Corus group of companies, please visit
http://www.corusgroup.com/entities

**********************************************************************




[Date Prev][Date Next]   [Thread Prev][Thread Next]   [Thread Index] [Date Index] [Author Index]