[Spacewalk-list] Taskomatic runs indefinitely without ever generating repodata
Matt Moldvan
matt at moldvan.com
Thu Jul 5 19:30:19 UTC 2018
Is there anything interesting
in /var/log/rhn/tasko/sat/channel-repodata-bunch? Do you have any hung
reposync processes? Any lingering Postgres locks that might be an issue?
It's odd that the run would only take 1 second, unless something is wrong
with the database or it's data...
What do you see from a spacewalk-sql command like below?
echo 'select label,name,modified,last_synced from rhnchannel' | sudo
spacewalk-sql -i
label | name |
modified | last_synced
----------------------------------+----------------------------------+-------------------------------+----------------------------
ovirt-x86_64-stable-6-nonprod | ovirt-x86_64-stable-6-nonprod |
2015-09-14 13:46:44.147134-05 |
extras7-x86_64-nonprod | extras7-x86_64-nonprod |
2017-11-06 10:26:30.011283-06 |
centos7-x86_64-all | centos7-x86_64-all |
2015-11-11 08:50:58.831234-06 | 2018-07-05 11:01:08.857-05
perl-5.16.x-all | perl-5.16.x-all |
2015-09-11 13:25:15.002198-05 | 2015-09-11 13:29:21.361-05
ovirt-x86_64-stable-6 | ovirt-x86_64-stable-6 |
2015-09-14 13:30:55.172-05 |
ovirt-x86_64-stable-6-prod | ovirt-x86_64-stable-6-prod |
2015-09-14 13:48:06.637063-05 |
other6-x86_64-all | other6-x86_64-all |
2015-07-28 09:20:38.156104-05 |
epel5-x86_64-all | epel5-x86_64-all |
2016-10-04 18:20:44.846312-05 | 2017-04-17 12:57:36.859-05
passenger6-x86_64-prod | passenger6-x86_64-prod |
2016-04-22 14:35:45.395518-05 |
perl-5.16.x-nonprod | perl-5.16.x-nonprod |
2015-09-11 13:27:32.261063-05 |
perl-5.16.x-prod | perl-5.16.x-prod |
2015-09-11 13:26:40.584715-05 | 2015-09-11 13:29:38.537-05
other6-x86_64-nonprod | other6-x86_64-nonprod |
2015-07-23 15:00:03.733479-05 |
other6-x86_64-prod | other6-x86_64-prod |
2015-07-21 15:10:48.719528-05 |
epel5-x86_64-prod | epel5-x86_64-prod |
2016-10-04 18:25:38.655383-05 |
passenger6-x86_64-all | passenger6-x86_64-all |
2016-04-20 11:37:19.002493-05 | 2016-04-20 11:58:42.312-05
docker7-x86_64-prod | docker7-x86_64-prod |
2017-08-03 11:42:08.474496-05 |
centos5-x86_64-nonprod | centos5-x86_64-nonprod |
2015-06-22 16:16:17.372799-05 |
other7-x86_64-nonprod | other7-x86_64-nonprod |
2016-07-14 13:03:10.320136-05 |
mongo3.2-centos6-x86_64-all | mongo3.2-centos6-x86_64-all |
2016-08-22 12:21:40.722182-05 | 2018-07-01 12:27:03.019-05
centos5-x86_64-prod | centos5-x86_64-prod |
2015-06-22 16:20:41.474486-05 |
passenger6-x86_64-nonprod | passenger6-x86_64-nonprod |
2016-04-20 12:29:24.677227-05 |
other7-x86_64-prod | other7-x86_64-prod |
2016-07-14 13:03:47.284295-05 |
cloudera5.7-x86_64-nonprod | cloudera5.7-x86_64-nonprod |
2016-05-09 12:10:16.496626-05 | 2016-06-20 13:11:20.62-05
epel5-x86_64-nonprod | epel5-x86_64-nonprod |
2016-10-04 18:25:09.844486-05 |
epel6-x86_64-prod | epel6-x86_64-prod |
2016-03-18 11:52:45.9199-05 | 2016-08-23 05:07:37.967-05
spacewalk6-client-all | spacewalk6-client-all |
2017-05-02 20:53:38.867018-05 | 2018-07-01 22:02:11.386-05
docker7-x86_64-nonprod | docker7-x86_64-nonprod |
2017-04-07 15:13:44.158973-05 |
mongo3.2-centos6-x86_64-nonprod | mongo3.2-centos6-x86_64-nonprod |
2016-08-22 12:34:18.095059-05 |
mongo3.2-centos6-x86_64-prod | mongo3.2-centos6-x86_64-prod |
2016-08-22 12:42:19.161165-05 |
local6-x86_64-all | local6-x86_64-all |
2015-09-30 08:55:37.657412-05 | 2016-04-19 07:00:23.632-05
centos5-x86_64-all | centos5-x86_64-all |
2015-06-22 15:20:22.085465-05 | 2017-04-17 13:09:39.635-05
spacewalk5-client-nonprod | spacewalk5-client-nonprod |
2017-05-02 20:53:20.430795-05 |
spacewalk5-client-prod | spacewalk5-client-prod |
2017-05-02 20:53:28.980968-05 |
spacewalk5-client-all | spacewalk5-client-all |
2017-05-02 20:53:08.276664-05 | 2018-07-05 10:10:11.665-05
spacewalk7-client-prod | spacewalk7-client-prod |
2017-05-02 20:54:32.321635-05 | 2018-07-05 11:01:14.499-05
epel6-x86_64-nonprod | epel6-x86_64-nonprod |
2016-03-18 11:52:14.915108-05 | 2018-07-05 10:10:08.774-05
centos7-x86_64-prod | centos7-x86_64-prod |
2015-11-11 09:02:06.69758-06 |
puppetlabs6-x86_64-prod | puppetlabs6-x86_64-prod |
2016-04-22 13:46:22.233841-05 | 2018-07-01 13:30:47.635-05
puppetlabs5-x86_64-nonprod | puppetlabs5-x86_64-nonprod |
2018-03-26 15:21:59.007749-05 | 2018-07-01 13:00:03.401-05
puppetlabs5-x86_64-prod | puppetlabs5-x86_64-prod |
2018-03-26 15:24:23.86552-05 | 2018-07-01 13:30:39.025-05
puppetlabs5-x86_64-all | puppetlabs5-x86_64-all |
2018-03-26 15:19:04.647981-05 | 2018-07-01 13:31:25.065-05
other5-x86_64-all | other5-x86_64-all |
2015-08-10 14:16:01.092867-05 |
other5-x86_64-nonprod | other5-x86_64-nonprod |
2015-08-10 14:18:05.114541-05 |
other5-x86_64-prod | other5-x86_64-prod |
2015-08-10 14:19:03.728982-05 |
centos6-x86_64-nonprod | centos6-x86_64-nonprod |
2015-06-22 16:24:07.137207-05 |
centos6-x86_64-prod | centos6-x86_64-prod |
2015-06-22 16:28:51.324002-05 |
extras7-x86_64-all | extras7-x86_64-all |
2017-08-16 09:13:26.8122-05 | 2018-07-05 10:05:10.626-05
centos6-x86_64-gitlab-ce-nonprod | centos6-x86_64-gitlab-ce-nonprod |
2017-04-17 11:43:36.609036-05 | 2018-07-05 10:04:57.277-05
spacewalk7-server-all | spacewalk7-server-all |
2017-03-28 15:22:31.851414-05 | 2018-07-05 11:11:31.564-05
local5-x86_64-all | local5-x86_64-all |
2016-02-24 12:19:36.791459-06 |
local5-x86_64-nonprod | local5-x86_64-nonprod |
2016-02-24 12:20:19.404008-06 |
local5-x86_64-prod | local5-x86_64-prod |
2016-02-24 12:20:45.098532-06 |
local6-x86_64-nonprod | local6-x86_64-nonprod |
2016-08-22 20:49:56.7376-05 |
local7-x86_64-all | local7-x86_64-all |
2016-07-14 13:00:32.511851-05 |
local7-x86_64-nonprod | local7-x86_64-nonprod |
2016-07-14 13:02:06.932169-05 |
local7-x86_64-prod | local7-x86_64-prod |
2016-07-14 13:02:38.496912-05 |
puppetlabs6-x86_64-all | puppetlabs6-x86_64-all |
2016-04-20 08:27:56.026914-05 | 2018-07-01 13:30:36.771-05
spacewalk7-client-nonprod | spacewalk7-client-nonprod |
2017-05-02 20:54:22.659512-05 | 2018-07-05 11:10:25.009-05
docker7-x86_64-all | docker7-x86_64-all |
2017-03-22 12:50:15.332561-05 | 2018-07-05 13:00:02.988-05
spacewalk7-client-all | spacewalk7-client-all |
2017-05-02 20:54:13.5076-05 | 2018-07-05 10:04:59.748-05
local6-x86_64-prod | local6-x86_64-prod |
2015-09-30 08:59:12.679727-05 |
centos6-x86_64-gitlab-ee-nonprod | centos6-x86_64-gitlab-ee-nonprod |
2016-04-14 11:39:01.432444-05 | 2018-07-05 11:12:20.525-05
mysqltools6-x86_64-all | mysqltools6-x86_64-all |
2016-03-17 12:41:37.44854-05 | 2018-07-05 12:00:02.319-05
mysqltools6-x86_64-nonprod | mysqltools6-x86_64-nonprod |
2016-03-17 12:58:35.036373-05 |
mysqltools6-x86_64-prod | mysqltools6-x86_64-prod |
2016-03-17 12:59:10.969162-05 |
spacewalk7-server-nonprod | spacewalk7-server-nonprod |
2017-03-28 15:23:02.210349-05 | 2018-07-05 11:12:47.471-05
spacewalk7-server-prod | spacewalk7-server-prod |
2017-03-28 15:23:29.309042-05 | 2017-05-02 20:56:45.247-05
epel7-x86_64-prod | epel7-x86_64-prod |
2016-03-22 09:48:38.060213-05 | 2018-07-05 09:57:25.861-05
puppetlabs6-x86_64-nonprod | puppetlabs6-x86_64-nonprod |
2016-04-20 12:28:55.337125-05 | 2018-07-01 13:30:43.362-05
newrelic-noarch-nover | newrelic-noarch-nover |
2016-10-13 13:54:38.621333-05 | 2016-10-13 14:09:41.778-05
other7-x86_64-all | other7-x86_64-all |
2016-07-14 13:01:25.848215-05 | 2018-07-05 14:00:03.714-05
spacewalk6-client-nonprod | spacewalk6-client-nonprod |
2017-05-02 20:53:50.507298-05 |
spacewalk6-client-prod | spacewalk6-client-prod |
2017-05-02 20:54:00.685324-05 |
spacewalk6-server-all | spacewalk6-server-all |
2018-06-22 23:11:30.637054-05 | 2018-07-05 11:01:11.543-05
puppetlabs7-x86_64-prod | puppetlabs7-x86_64-prod |
2016-07-14 13:29:04.67033-05 | 2018-07-01 13:31:29.425-05
spacewalk6-server-nonprod | spacewalk6-server-nonprod |
2018-06-22 23:17:20.660409-05 |
spacewalk6-server-prod | spacewalk6-server-prod |
2018-06-22 23:18:02.738869-05 |
puppetlabs7-x86_64-nonprod | puppetlabs7-x86_64-nonprod |
2016-07-14 13:28:34.475051-05 | 2018-07-01 13:16:25.948-05
epel6-x86_64-all | epel6-x86_64-all |
2016-03-18 11:50:17.587171-05 | 2018-07-05 11:07:42.644-05
centos6-x86_64-gitlab-ee | centos6-x86_64-gitlab-ee |
2015-12-24 13:21:10.493684-06 | 2018-07-05 11:08:30.039-05
puppetlabs7-x86_64-all | puppetlabs7-x86_64-all |
2016-07-14 12:54:59.388232-05 | 2018-07-01 13:32:02.745-05
epel7-x86_64-nonprod | epel7-x86_64-nonprod |
2016-03-22 09:47:34.668867-05 | 2017-04-21 11:08:24.573-05
centos6-x86_64-all | centos6-x86_64-all |
2015-06-22 15:19:13.053429-05 | 2018-07-02 01:12:57.768-05
epel7-x86_64-all | epel7-x86_64-all |
2016-03-22 09:44:48.748142-05 | 2018-07-05 09:11:28.553-05
centos7-x86_64-nonprod | centos7-x86_64-nonprod |
2015-10-21 22:02:28.107902-05 |
(85 rows)
On Thu, Jul 5, 2018 at 11:48 AM Gerald Vogt <vogt at spamcop.net> wrote:
> On 05.07.18 16:05, Matt Moldvan wrote:
> > How is the server utilization with respect to disk I/O (something like
> > iotop or htop might help here)? Maybe there is something else blocking
>
> My server is basically idle. 99% idle, little disk i/o. It doesn't do
> anything really.
>
> > and the server doesn't have enough resources to complete. Have you
> > tried running an strace against the running process?
>
> If it doesn't have enough resources shouldn't there be an exception?
>
> For me, it looks more like something doesn't make it into the database and
> thus into the persistent state. For instance, I now have the repodata task
> at "RUNNING" for three days:
>
> Channel Repodata: 2018-07-02 08:13:10 CEST RUNNING
>
> The log file shows this regarding repodata:
>
> > # fgrep -i repodata rhn_taskomatic_daemon.log
> > INFO | jvm 1 | 2018/07/02 08:13:10 | 2018-07-02 08:13:10,584
> [Thread-12] INFO com.redhat.rhn.taskomatic.TaskoQuartzHelper - Job
> single-channel-repodata-bunch-0 scheduled succesfully.
> > INFO | jvm 1 | 2018/07/02 08:13:10 | 2018-07-02 08:13:10,636
> [DefaultQuartzScheduler_Worker-8] INFO com.redhat.rhn.taskomatic.TaskoJob
> - single-channel-repodata-bunch-0: bunch channel-repodata-bunch STARTED
> > INFO | jvm 1 | 2018/07/02 08:13:10 | 2018-07-02 08:13:10,651
> [DefaultQuartzScheduler_Worker-8] DEBUG com.redhat.rhn.taskomatic.TaskoJob
> - single-channel-repodata-bunch-0: task channel-repodata started
> > INFO | jvm 1 | 2018/07/02 08:13:10 | 2018-07-02 08:13:10,793
> [DefaultQuartzScheduler_Worker-8] INFO
> com.redhat.rhn.taskomatic.task.ChannelRepodata - In the queue: 4
> > INFO | jvm 1 | 2018/07/02 08:13:11 | 2018-07-02 08:13:11,102
> [DefaultQuartzScheduler_Worker-8] DEBUG com.redhat.rhn.taskomatic.TaskoJob
> - channel-repodata (single-channel-repodata-bunch-0) ... running
> > INFO | jvm 1 | 2018/07/02 08:13:11 | 2018-07-02 08:13:11,103
> [DefaultQuartzScheduler_Worker-8] INFO com.redhat.rhn.taskomatic.TaskoJob
> - single-channel-repodata-bunch-0: bunch channel-repodata-bunch FINISHED
>
> So according to the logs the repodata bunch has finished. According to the
> web interface it has not. Nothing has been updated in
> /var/cache/rhn/repodata/ either. In addition, those four channels which
> were still updated haven't been updated either now.
>
> Thanks,
>
> Gerald
>
>
>
> >
> > I also had an (well, many) issue(s) with our Spacewalk server before
> > disabling snapshots in /etc/rhn/rhn.conf. I also increased the number
> > of workers and max repodata work items:
> >
> > # system snapshots enabled
> > enable_snapshots = 0
> > ...
> > taskomatic.maxmemory=6144
> > taskomatic.errata_cache_max_work_items = 500
> > taskomatic.channel_repodata_max_work_items = 50
> > taskomatic.channel_repodata_workers = 5
> >
> >
> >
> > On Thu, Jul 5, 2018 at 4:38 AM Florence Savary
> > <florence.savary.fs at gmail.com <mailto:florence.savary.fs at gmail.com>>
> wrote:
> >
> > Hello,
> >
> > Thanks for sharing your configuration files. They differ very little
> > from mine. I just changed the number of workers in rhn.conf, but it
> > didn't change anything.
> >
> > I deleted all the channels clones not used by any system and dating
> > back from before May 2018, in order to lower the number of channels
> > in the queue. There were 127 channels in the queue before these
> > deletion (indicated in /var/log/rhn/rhn_taskomatic_daemon.log), and
> > there are 361 of them now ... I must admit I'm confused... I hoped
> > it would reduce the number of channels to process and thus "help"
> > taskomatic, but obviously I was wrong.
> >
> > I also noticed that the repodata regeneration seems to work fine for
> > existing channels that are not clones, but it is not working for new
> > channels that are not clones (and not working for new clones but
> > nothing new here).
> >
> > Has anyone got any other idea (even the tiniest) ?
> >
> > Regards,
> > Florence
> >
> >
> > 2018-07-04 15:21 GMT+02:00 Paul Dias - BCX <paul.dias at bcx.co.za
> > <mailto:paul.dias at bcx.co.za>>:
> >
> > Hi,____
> >
> > __ __
> >
> > Let me post my settings that I have on my CentOS6 server. Can’t
> > remember but I have one or two others, but his is from the top
> > of my head.____
> >
> > __ __
> >
> > /etc/rhn/rhn.conf____
> >
> > # Added by paul dias increase number of taskomatic workers
> > 20180620____
> >
> > taskomatic.channel_repodata_workers = 3____
> >
> > taskomatic.java.maxmemory=4096____
> >
> > __ __
> >
> > /etc/sysconfig/tomcat6____
> >
> > JAVA_OPTS="-ea -Xms256m -Xmx512m -Djava.awt.headless=true
> > -Dorg.xml.sax.driver=org.apache.xerces.parsers.SAXParser
> > -Dorg.apache.tomcat.util.http.Parameters.MAX_COUNT=1024
> > -XX:MaxNewSize=256 -XX:-UseConcMarkSweepGC
> > -Dnet.sf.ehcache.skipUpdateCheck=true
> >
> -Djavax.sql.DataSource.Factory=org.apache.commons.dbcp.BasicDataSourceFactory"____
> >
> > __ __
> >
> > /etc/tomcat/server.xml____
> >
> > <!-- Define an AJP 1.3 Connector on port 8009 -->____
> >
> > <Connector port="8009" protocol="AJP/1.3"
> > redirectPort="8443" URIEncoding="UTF-8" address="127.0.0.1"
> > maxThreads="256" connectionTimeout="20000"/>____
> >
> > __ __
> >
> > <Connector port="8009" protocol="AJP/1.3"
> > redirectPort="8443" URIEncoding="UTF-8" address="::1"
> > maxThreads="256" connectionTimeout="20000"/>____
> >
> > __ __
> >
> > /usr/share/rhn/config-defaults/rhn_taskomatic_daemon.conf____
> >
> > # Initial Java Heap Size (in MB)____
> >
> > wrapper.java.initmemory=512____
> >
> > __ __
> >
> > # Maximum Java Heap Size (in MB)____
> >
> > wrapper.java.maxmemory=1512____
> >
> > # Adjusted by paul 20180620____
> >
> > __ __
> >
> > wrapper.ping.timeout=0____
> >
> > # # adjusted paul dias 20180620____
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > Regards,____
> >
> > *Paul Dias____*
> >
> > Technical Consultant____
> >
> > 6^th Floor, 8 Boundary Road____
> >
> > Newlands____
> >
> > Cape Town____
> >
> > 7700____
> >
> > T: +27 (0) 21 681 3149 <+27%2021%20681%203149>
> <tel:+27%2021%20681%203149>____
> >
> > *Meet your future today.____*
> >
> > *__ __*
> >
> > __BCX______
> >
> > __ __
> >
> > __ __
> >
> > __ __
> >
> > __Social-facebook
> > <https://www.facebook.com/BCXworld>____Social-twitter
> > <https://twitter.com/BCXworld>____Social-linkdin
> > <https://za.linkedin.com/BCX>____Social-youtube
> > <https://www.youtube.com/BCXworld>______
> >
> > __ __
> >
> > __ __
> >
> > This e-mail is subject to the BCX electronic communication legal
> > notice, available at:
> > https://www.bcx.co.za/disclaimers____
> >
> > /__ __/
> >
> > /__ __/
> >
> > __ __
> >
> > *From:*Paul Dias - BCX
> > *Sent:* 02 July 2018 06:53 PM
> >
> >
> > *To:* spacewalk-list at redhat.com <mailto:
> spacewalk-list at redhat.com>
> > *Subject:* Re: [Spacewalk-list] Taskomatic runs indefinitely
> > without ever generating repodata____
> >
> > __ __
> >
> > What I have noticed, if you use
> > "spacecmd softchannel_generateyumcache <channel name>" and then
> > go to tasks and run single repodata bunch, you will see it will
> > actually start and generate your channel cache for you on the
> > channel you used the spacecmd on, this works every time.____
> >
> > __ __
> >
> > But yes the task logs just show repodata bunch running
> forever.____
> >
> > __ __
> >
> > Regards,____
> >
> > *Paul Dias*____
> >
> > 6^th Floor, 8 Boundary Road____
> >
> > Newlands____
> >
> > Cape Town____
> >
> > 7700____
> >
> > T: +27 (0) 21 681 3149 <+27%2021%20681%203149>
> <tel:+27%2021%20681%203149>____
> >
> > __ __
> >
> > *Meet your future today.*____
> >
> > **____
> >
> > BCX____
> >
> > __ __
> >
> >
> ------------------------------------------------------------------------
> >
> > *From:*Gerald Vogt <vogt at spamcop.net <mailto:vogt at spamcop.net>>
> > *Sent:* Monday, 02 July 2018 9:45 AM
> > *To:* spacewalk-list at redhat.com <mailto:
> spacewalk-list at redhat.com>
> > *Subject:* Re: [Spacewalk-list] Taskomatic runs indefinitely
> > without ever generating repodata____
> >
> > ____
> >
> > After letting the upgraded server sit for a while it seems only
> > a few of
> > the task schedules actually finish. By now, only those tasks
> > show up in
> > in the task engine status page:
> >
> > Changelog Cleanup: 2018-07-01 23:00:00 CEST
> FINISHED
> > Clean Log History: 2018-07-01 23:00:00 CEST
> FINISHED
> > Compare Config Files: 2018-07-01 23:00:00 CEST
> FINISHED
> > Daily Summary Mail: 2018-07-01 23:00:00 CEST
> FINISHED
> > Daily Summary Queue: 2018-07-01 23:00:00 CEST
> FINISHED
> >
> > All the other tasks have disappeared from the list by now.
> >
> > The repo-sync tasks seem to work. New packages appear in the
> > channel.
> > However, the repo build is not running or better it seems to
> never
> > properly finish.
> >
> > If I start it manually, it seems to do its work:
> >
> > > INFO | jvm 1 | 2018/07/02 08:13:10 | 2018-07-02
> 08:13:10,584 [Thread-12] INFO com.redhat.rhn.taskomatic.TaskoQuartzHelper
> - Job single-channel-repodata-bunch-0 scheduled succesfully.
> > > INFO | jvm 1 | 2018/07/02 08:13:10 | 2018-07-02
> 08:13:10,636 [DefaultQuartzScheduler_Worker-8] INFO
> com.redhat.rhn.taskomatic.TaskoJob - single-channel-repodata-bunch-0: bunch
> channel-repodata-bunch STARTED
> > > INFO | jvm 1 | 2018/07/02 08:13:10 | 2018-07-02
> 08:13:10,651 [DefaultQuartzScheduler_Worker-8] DEBUG
> com.redhat.rhn.taskomatic.TaskoJob - single-channel-repodata-bunch-0: task
> channel-repodata started
> > > INFO | jvm 1 | 2018/07/02 08:13:10 | 2018-07-02
> 08:13:10,793 [DefaultQuartzScheduler_Worker-8] INFO
> com.redhat.rhn.taskomatic.task.ChannelRepodata - In the queue: 4
> > > INFO | jvm 1 | 2018/07/02 08:13:11 | 2018-07-02
> 08:13:11,102 [DefaultQuartzScheduler_Worker-8] DEBUG
> com.redhat.rhn.taskomatic.TaskoJob - channel-repodata
> (single-channel-repodata-bunch-0) ... running
> > > INFO | jvm 1 | 2018/07/02 08:13:11 | 2018-07-02
> 08:13:11,103 [DefaultQuartzScheduler_Worker-8] INFO
> com.redhat.rhn.taskomatic.TaskoJob - single-channel-repodata-bunch-0: bunch
> channel-repodata-bunch FINISHED
> > > INFO | jvm 1 | 2018/07/02 08:13:11 | 2018-07-02
> 08:13:11,137 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - File Modified
> Date:2018-06-23 03:48:50 CEST
> > > INFO | jvm 1 | 2018/07/02 08:13:11 | 2018-07-02
> 08:13:11,137 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Channel Modified
> Date:2018-07-02 03:45:39 CEST
> > > INFO | jvm 1 | 2018/07/02 08:13:11 | 2018-07-02
> 08:13:11,211 [Thread-678] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - File Modified
> Date:2018-06-23 04:09:51 CEST
> > > INFO | jvm 1 | 2018/07/02 08:13:11 | 2018-07-02
> 08:13:11,213 [Thread-678] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Channel Modified
> Date:2018-07-02 03:47:55 CEST
> > > INFO | jvm 1 | 2018/07/02 08:13:19 | 2018-07-02
> 08:13:19,062 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Generating new
> repository metadata for channel 'epel6-centos6-x86_64'(sha1) 14401
> packages, 11613 errata
> > > INFO | jvm 1 | 2018/07/02 08:13:21 | 2018-07-02
> 08:13:21,193 [Thread-678] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Generating new
> repository metadata for channel 'epel7-centos7-x86_64'(sha1) 16282
> packages, 10176 errata
> > > INFO | jvm 1 | 2018/07/02 08:40:12 | 2018-07-02
> 08:40:12,351 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Repository
> metadata generation for 'epel6-centos6-x86_64' finished in 1613 seconds
> > > INFO | jvm 1 | 2018/07/02 08:40:12 | 2018-07-02
> 08:40:12,457 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - File Modified
> Date:2018-06-19 06:28:57 CEST
> > > INFO | jvm 1 | 2018/07/02 08:40:12 | 2018-07-02
> 08:40:12,457 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Channel Modified
> Date:2018-07-02 04:30:05 CEST
> > > INFO | jvm 1 | 2018/07/02 08:40:12 | 2018-07-02
> 08:40:12,691 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Generating new
> repository metadata for channel 'postgresql96-centos7-x86_64'(sha256) 1032
> packages, 0 errata
> > > INFO | jvm 1 | 2018/07/02 08:41:51 | 2018-07-02
> 08:41:51,710 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Repository
> metadata generation for 'postgresql96-centos7-x86_64' finished in 98 seconds
> > > INFO | jvm 1 | 2018/07/02 08:41:51 | 2018-07-02
> 08:41:51,803 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - File Modified
> Date:2018-06-20 05:08:38 CEST
> > > INFO | jvm 1 | 2018/07/02 08:41:51 | 2018-07-02
> 08:41:51,803 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Channel Modified
> Date:2018-07-02 04:00:00 CEST
> > > INFO | jvm 1 | 2018/07/02 08:41:51 | 2018-07-02
> 08:41:51,923 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Generating new
> repository metadata for channel 'postgresql10-centos6-x86_64'(sha512) 436
> packages, 0 errata
> > > INFO | jvm 1 | 2018/07/02 08:42:26 | 2018-07-02
> 08:42:26,479 [Thread-677] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Repository
> metadata generation for 'postgresql10-centos6-x86_64' finished in 34 seconds
> > > INFO | jvm 1 | 2018/07/02 08:45:01 | 2018-07-02
> 08:45:01,697 [Thread-678] INFO
> com.redhat.rhn.taskomatic.task.repomd.RepositoryWriter - Repository
> metadata generation for 'epel7-centos7-x86_64' finished in 1900 seconds
> >
> > yet, the task remains in RUNNING. And for whatever reason it
> > only seems
> > to work some channels. I find a total of 20 repos syncing in the
> > logs of
> > the updated server compared to 42 repos syncing in the logs of
> > the old.
> > I don't really see the difference between those 20 repos syncing
> > and
> > those other 22 not. First I suspected channels with custom quartz
> > schedules, but then I found channels in both groups.
> >
> > So I don't know how to troubleshoot this any further. The
> > repodata task
> > which I have started 1,5 hours ago is still at "RUNNING". The
> > channels
> > for which the sync works have been updated. I don't know why it
> > is still
> > running. Server load is back down...
> >
> > Thanks,
> >
> > Gerald
> >
> > On 22.06.18 19:12, Gerald Vogt wrote:
> > > I have the same problem after upgrading from 2.6 to 2.8 on
> CentOS 6.9. I
> > > have even increased the memory as suggested by that link but
> it makes no
> > > differences. None of the scheduled tasks are running. I can
> run a bunch
> > > manually. But the scheduler doesn't seem to work. Last
> execution times
> > > on the task engine status pages are still at timestamps from
> before the
> > > upgrade. -Gerald
> > >
> > >
> > >
> > > On 22.06.18 14:15, Avi Miller wrote:
> > >> Hi,
> > >>
> > >>> On 22 Jun 2018, at 5:51 pm, Florence Savary
> > >>> <florence.savary.fs at gmail.com
> > <mailto:florence.savary.fs at gmail.com>> wrote:
> > >>>
> > >>> When using taskotop, we can see a line for the
> channel-repodata task,
> > >>> we see it is running, but there is never any channel
> displayed in the
> > >>> Channel column. We can also see the task marked as running
> in the
> > >>> Admin tab of the WebUI, but if we let it, it never stops.
> The task
> > >>> runs indefinitely, whithout ever doing anything.
> > >>
> > >> If you've never modified the default memory settings,
> Taskomatic is
> > >> probably running out of memory and task is crashing. This is
> a known
> > >> issue, particularly when you sync large repos.
> > >>
> > >> I would suggest increasing the memory assigned to Taskomatic
> to see if
> > >> that resolves the issue. You will need to restart it after
> making
> > >> these changes:
> > >>
> https://docs.oracle.com/cd/E92593_01/E90695/html/swk24-issues-memory.html
> > >>
> > >> Cheers,
> > >> Avi
> > >>
> > >> --
> > >> Oracle <http://www.oracle.com>
> > >> Avi Miller | Product Management Director | +61 (3) 8616 3496
> <+61%203%208616%203496> <tel:+61%203%208616%203496>
> > >> Oracle Linux and Virtualization
> > >> 417 St Kilda Road, Melbourne, Victoria 3004 Australia
> > >>
> > >>
> > >> _______________________________________________
> > >> Spacewalk-list mailing list
> > >> Spacewalk-list at redhat.com <mailto:Spacewalk-list at redhat.com>
> > >> https://www.redhat.com/mailman/listinfo/spacewalk-list
> > >>
> > >
> > > _______________________________________________
> > > Spacewalk-list mailing list
> > > Spacewalk-list at redhat.com <mailto:Spacewalk-list at redhat.com>
> > > https://www.redhat.com/mailman/listinfo/spacewalk-list
> >
> > ____
> >
> >
> > _______________________________________________
> > Spacewalk-list mailing list
> > Spacewalk-list at redhat.com <mailto:Spacewalk-list at redhat.com>
> > https://www.redhat.com/mailman/listinfo/spacewalk-list
> >
> >
> > _______________________________________________
> > Spacewalk-list mailing list
> > Spacewalk-list at redhat.com <mailto:Spacewalk-list at redhat.com>
> > https://www.redhat.com/mailman/listinfo/spacewalk-list
> >
> >
> >
> > _______________________________________________
> > Spacewalk-list mailing list
> > Spacewalk-list at redhat.com
> > https://www.redhat.com/mailman/listinfo/spacewalk-list
> >
>
> _______________________________________________
> Spacewalk-list mailing list
> Spacewalk-list at redhat.com
> https://www.redhat.com/mailman/listinfo/spacewalk-list
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://listman.redhat.com/archives/spacewalk-list/attachments/20180705/51ec664d/attachment.htm>
More information about the Spacewalk-list
mailing list