[389-users] 389 unusable on F11?

Kevin Bowling kevinb at analograils.com
Fri Sep 11 17:14:01 UTC 2009


On 9/11/2009 12:43 PM, Noriko Hosoi wrote:
> On 09/10/2009 07:46 PM, Kevin Bowling wrote:
>> Hi,
>>
>> I have been running FDS/389 on a F11 xen DomU for several months.  I 
>> use it as the backend for UNIX username/passwords and also for 
>> redMine (a Ruby on Rails bug tracker) for http://www.gnucapplus.org/.
>>
>> This VM would regularly lock up every week or so when 389 was still 
>> called FDS.  I've since upgraded to 389 by issuing 'yum upgrade' as 
>> well as running the 'setup-...-.pl -u' script and now it barely goes 
>> a day before crashing.  When ldap crashes, the whole box basically 
>> becomes unresponsive.
>>
>> I left the Xen hardware console open to see what was up and the only 
>> thing I could conclude was that 389 was crashing (if I issued a 
>> service start it came back to life).  Doing anything like a top or ls 
>> will completely kill the box.  Likewise, the logs show nothing at or 
>> before the time of crash.  I suspected too few file descriptors but 
>> changing that to a very high number had no impact.
>>
>> I was about to do a rip and replace with OpenLDAP which I use very 
>> sucesessfully for our corporate systems but figured I ought to see if 
>> anyone here can help or if I can submit any kind of meaningful bug 
>> report first.  I assume I will need to run 389's slapd without 
>> daemonizing it and hope it spits something useful out to stderr.  Any 
>> advice here would be greatly appreciated, as would any success 
>> stories of using 389 on F11.
> Hello Kevin,
>
> You specified the platform "F11 xen DomU".  Did you have a chance to 
> run the 389 server on any other platforms?  I'm wondering if the crash 
> is observed only on the specific platform or not.  Is the server 
> running on the 64-bit machine or 32-bit?
>
> If you start the server with "-d 1" option, the server will run as the 
> trace mode.  (E.g., /usr/lib[64]/dirsrv/slapd-YOURID/start-slapd -d 1)
>
> I'm afraid it might be a memory leak.  When you restart the 389 
> server, could you check the size of ns-slapd some time like every hour 
> and see if the server size keeps growing or stops?  Also, the server 
> quits if it fails to write to the errors log.  If it happens, it's 
> logged in the system log.  Does the messages file on the system  
> happen to have some logs related to the 389 server?
>
> Thanks,
> --noriko
>>
>> I'm not subscribed to the list so please CC.
>>
>> Regards,
>>
>> Kevin Bowing
>>
>> -- 
>> 389 users mailing list
>> 389-users at redhat.com
>> https://www.redhat.com/mailman/listinfo/fedora-directory-users
>

I captured some output while running in trace, see the end of this 
message.  The system is 64-bit, I have not run on any other boxes.  A 
cursory look with top showed only 10MB or so RSS memory.

Regards,
Kevin

[11/Sep/2009:09:58:44 -0700] - => id2entry( 48 )
[11/Sep/2009:09:58:44 -0700] - <= id2entry 7f025401f5a0 (cache)
[11/Sep/2009:09:58:44 -0700] - => id2entry( 50 )
[11/Sep/2009:09:58:44 -0700] - <= id2entry 7f0254021190 (cache)
[11/Sep/2009:09:58:44 -0700] - => slapi_reslimit_get_integer_limit() 
conn=0xa856beb0, handle=3
[11/Sep/2009:09:58:44 -0700] - <= slapi_reslimit_get_integer_limit() 
returning NO VALUE
[11/Sep/2009:09:58:44 -0700] - => slapi_reslimit_get_integer_limit() 
conn=0xa856bc60, handle=3
[11/Sep/2009:09:58:44 -0700] - <= slapi_reslimit_get_integer_limit() 
returning NO VALUE
[11/Sep/2009:09:58:44 -0700] - => slapi_reslimit_get_integer_limit() 
conn=0xa856bd88, handle=3
[11/Sep/2009:09:58:44 -0700] - <= slapi_reslimit_get_integer_limit() 
returning NO VALUE
[11/Sep/2009:09:58:44 -0700] - => slapi_reslimit_get_integer_limit() 
conn=0xa856bb38, handle=3
[11/Sep/2009:09:58:44 -0700] - <= slapi_reslimit_get_integer_limit() 
returning NO VALUE
[11/Sep/2009:09:58:44 -0700] - => slapi_reslimit_get_integer_limit() 
conn=0xa856ba10, handle=3
[11/Sep/2009:09:58:44 -0700] - <= slapi_reslimit_get_integer_limit() 
returning NO VALUE
[11/Sep/2009:09:58:44 -0700] - => send_ldap_result 0::
[11/Sep/2009:09:58:44 -0700] - <= send_ldap_result
[11/Sep/2009:09:58:50 -0700] - ldbm backend flushing
[11/Sep/2009:09:58:50 -0700] - ldbm backend done flushing
[11/Sep/2009:09:58:50 -0700] - ldbm backend flushing
[11/Sep/2009:09:58:50 -0700] - ldbm backend done flushing
[11/Sep/2009:09:59:20 -0700] - ldbm backend flushing
[11/Sep/2009:09:59:20 -0700] - ldbm backend done flushing
[11/Sep/2009:09:59:20 -0700] - ldbm backend flushing
[11/Sep/2009:09:59:20 -0700] - ldbm backend done flushing
[11/Sep/2009:09:59:50 -0700] - ldbm backend flushing
[11/Sep/2009:09:59:50 -0700] - ldbm backend done flushing
[11/Sep/2009:09:59:50 -0700] - ldbm backend flushing
[11/Sep/2009:09:59:50 -0700] - ldbm backend done flushing
[11/Sep/2009:10:00:20 -0700] - ldbm backend flushing
[11/Sep/2009:10:00:20 -0700] - ldbm backend done flushing
[11/Sep/2009:10:00:20 -0700] - ldbm backend flushing
[11/Sep/2009:10:00:20 -0700] - ldbm backend done flushing
[11/Sep/2009:10:00:50 -0700] - ldbm backend flushing
[11/Sep/2009:10:01:03 -0700] - ldbm backend done flushing
[11/Sep/2009:10:01:03 -0700] - ldbm backend flushing
[11/Sep/2009:10:01:04 -0700] - ldbm backend done flushing
[11/Sep/2009:10:01:35 -0700] - ldbm backend flushing
[11/Sep/2009:10:01:39 -0700] - ldbm backend done flushing
[11/Sep/2009:10:01:39 -0700] - ldbm backend flushing
[11/Sep/2009:10:01:39 -0700] - ldbm backend done flushing
[11/Sep/2009:10:01:39 -0700] - ldbm backend flushing
[11/Sep/2009:10:01:39 -0700] - ldbm backend done flushing
[11/Sep/2009:10:01:39 -0700] - ldbm backend flushing
[11/Sep/2009:10:01:39 -0700] - ldbm backend done flushing
[11/Sep/2009:10:01:39 -0700] - slapd shutting down - signaling operation 
threads
[11/Sep/2009:10:01:40 -0700] - slapd shutting down - waiting for 30 
threads to terminate
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - slapd shutting down - waiting for 29 
threads to terminate
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:40 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - op_thread received shutdown signal
[11/Sep/2009:10:01:41 -0700] - slapd shutting down - waiting for 28 
threads to terminate
[11/Sep/2009:10:01:41 -0700] - slapd shutting down - closing down 
internal subsystems and plugins
[11/Sep/2009:10:01:41 -0700] - slapd shutting down - waiting for 
backends to close down
[11/Sep/2009:10:01:42 -0700] - => slapi_control_present (looking for 
1.3.6.1.4.1.42.2.27.8.5.1)
[11/Sep/2009:10:01:42 -0700] - <= slapi_control_present 0 (NO CONTROLS)
[11/Sep/2009:10:01:42 -0700] - modify_update_last_modified_attr
[11/Sep/2009:10:01:42 -0700] - Calling plugin 'Distributed Numeric 
Assignment internal preop plugin' #0 type 421
[11/Sep/2009:10:01:42 -0700] dna-plugin - --> dna_pre_op
[11/Sep/2009:10:01:42 -0700] dna-plugin - <-- dna_pre_op
[11/Sep/2009:10:01:42 -0700] - Calling plugin 'Legacy replication 
internal preoperation plugin' #1 type 421
[11/Sep/2009:10:01:42 -0700] - Calling plugin 'Multimaster replication 
internal preoperation plugin' #2 type 421
[11/Sep/2009:10:01:42 -0700] - => entry_apply_mods
[11/Sep/2009:10:01:42 -0700] - <= entry_apply_mods 0
[11/Sep/2009:10:01:42 -0700] - => send_ldap_result 0::
[11/Sep/2009:10:01:42 -0700] - <= send_ldap_result
[11/Sep/2009:10:01:42 -0700] - ps_service_persistent_searches: entry 
"cn=uniqueid generator,cn=config" not enqueued on any persistent search 
lists
[11/Sep/2009:10:01:42 -0700] - Calling plugin 'Class of Service 
internalpostoperation plugin' #0 type 521
[11/Sep/2009:10:01:42 -0700] - --> cos_post_op
[11/Sep/2009:10:01:42 -0700] - --> cos_cache_change_notify
[11/Sep/2009:10:01:42 -0700] - --> cos_cache_template_index_bsearch
[11/Sep/2009:10:01:42 -0700] - --> cos_cache_getref
[11/Sep/2009:10:01:42 -0700] - <-- cos_cache_getref
[11/Sep/2009:10:01:42 -0700] - <-- cos_cache_template_index_bsearch
[11/Sep/2009:10:01:42 -0700] - <-- cos_cache_change_notify
[11/Sep/2009:10:01:42 -0700] - <-- cos_post_op
[11/Sep/2009:10:01:42 -0700] - Calling plugin 'Legacy replication 
internal postoperation plugin' #1 type 521
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'Multimaster replication 
internal postoperation plugin' #2 type 521
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'Retrocl internal 
postoperation plugin' #3 type 521
not applying change if not logging
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'Roles 
internalpostoperation plugin' #4 type 521
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'Legacy Replication 
Plugin' #0 type 210
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'Roles Plugin' #0 type 210
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'Multimaster Replication 
Plugin' #0 type 210
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'HTTP Client' #0 type 210
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'Class of Service' #0 type 210
[11/Sep/2009:10:01:43 -0700] - --> cos_close
[11/Sep/2009:10:01:43 -0700] - --> cos_cache_stop
[11/Sep/2009:10:01:43 -0700] - <-- cos_cache_wait_on_change thread exit
[11/Sep/2009:10:01:43 -0700] - --> cos_cache_release
[11/Sep/2009:10:01:43 -0700] - <-- cos_cache_release
[11/Sep/2009:10:01:43 -0700] - <-- cos_cache_stop
[11/Sep/2009:10:01:43 -0700] - <-- cos_close
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'ACL Plugin' #0 type 210
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'Views' #0 type 210
[11/Sep/2009:10:01:43 -0700] views-plugin - --> views_close
[11/Sep/2009:10:01:43 -0700] views-plugin - --> views_cache_free
[11/Sep/2009:10:01:43 -0700] views-plugin - <-- views_cache_free
[11/Sep/2009:10:01:43 -0700] views-plugin - <-- views_close
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'State Change Plugin' #0 
type 210
[11/Sep/2009:10:01:43 -0700] statechange-plugin - --> statechange_close
[11/Sep/2009:10:01:43 -0700] statechange-plugin - <-- statechange_close
[11/Sep/2009:10:01:43 -0700] - Calling plugin 'ldbm database' #0 type 210
[11/Sep/2009:10:01:43 -0700] - ldbm backend syncing
[11/Sep/2009:10:01:43 -0700] - Waiting for 4 database threads to stop
[11/Sep/2009:10:01:43 -0700] - Leaving deadlock_threadmain
[11/Sep/2009:10:01:44 -0700] - Leaving checkpoint_threadmain before 
checkpoint
[11/Sep/2009:10:01:44 -0700] - Checkpointing database ...
[11/Sep/2009:10:01:44 -0700] - Leaving checkpoint_threadmain
[11/Sep/2009:10:01:44 -0700] - Leaving trickle_threadmain priv
[11/Sep/2009:10:01:44 -0700] - Leaving perf_threadmain
[11/Sep/2009:10:01:45 -0700] - All database threads now stopped
[11/Sep/2009:10:01:45 -0700] - ldbm backend done syncing
[11/Sep/2009:10:01:45 -0700] - Calling plugin 'chaining database' #0 
type 210
[11/Sep/2009:10:01:45 -0700] - Removed [1] entries from the dse tree.
[11/Sep/2009:10:01:45 -0700] - Removed [166] entries from the dse tree.
[11/Sep/2009:10:01:45 -0700] - ldbm backend cleaning up
[11/Sep/2009:10:01:45 -0700] - ldbm backend cleaning up
[11/Sep/2009:10:01:45 -0700] - slapd shutting down - backends closed down
[11/Sep/2009:10:01:45 -0700] - => reslimit_update_from_entry() 
conn=0xa856ba10, entry=0x0
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 0 (based on nsLookThroughLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 1 (based on nsSizeLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 2 (based on nsTimeLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 3 (based on nsIdleTimeout)
[11/Sep/2009:10:01:45 -0700] - <= reslimit_update_from_entry() returning 
status 0
[11/Sep/2009:10:01:45 -0700] - => reslimit_update_from_entry() 
conn=0xa856bb38, entry=0x0
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 0 (based on nsLookThroughLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 1 (based on nsSizeLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 2 (based on nsTimeLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 3 (based on nsIdleTimeout)
[11/Sep/2009:10:01:45 -0700] - <= reslimit_update_from_entry() returning 
status 0
[11/Sep/2009:10:01:45 -0700] - => reslimit_update_from_entry() 
conn=0xa856bd88, entry=0x0
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 0 (based on nsLookThroughLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 1 (based on nsSizeLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 2 (based on nsTimeLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 3 (based on nsIdleTimeout)
[11/Sep/2009:10:01:45 -0700] - <= reslimit_update_from_entry() returning 
status 0
[11/Sep/2009:10:01:45 -0700] - => reslimit_update_from_entry() 
conn=0xa856beb0, entry=0x0
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 0 (based on nsLookThroughLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 1 (based on nsSizeLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 2 (based on nsTimeLimit)
[11/Sep/2009:10:01:45 -0700] - reslimit_update_from_entry(): setting 
limit for handle 3 (based on nsIdleTimeout)
[11/Sep/2009:10:01:45 -0700] - <= reslimit_update_from_entry() returning 
status 0
[11/Sep/2009:10:01:45 -0700] - slapd stopped.




More information about the Fedora-directory-users mailing list