<div dir="ltr"><div dir="ltr"><div>Hi,</div><div><br></div><div>Starting with 3.2 you should probably start with the minimal arch <a href="https://softwarefactory-project.io/cgit/software-factory/sf-config/tree/refarch/minimal.yaml">https://softwarefactory-project.io/cgit/software-factory/sf-config/tree/refarch/minimal.yaml</a></div><div>Then add others components you need step by step.<br></div><div><br></div><div>Regards,</div><div>Fabien<br></div></div></div><br><div class="gmail_quote"><div dir="ltr" class="gmail_attr">On Fri, Apr 12, 2019 at 12:26 PM Tristan Cacqueray <<a href="mailto:tdecacqu@redhat.com">tdecacqu@redhat.com</a>> wrote:<br></div><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left:1px solid rgb(204,204,204);padding-left:1ex"><br>
On Fri, Apr 12, 2019 at 09:18 RUIZ LOPEZ Noel wrote:<br>
> Hello<br>
><br>
> First of all, thanks for your quick answer.<br>
><br>
> I have tried to upgrade from 3.0 to 3.2, but I get this error when I try to install the 3.2<br>
> "Error: centos-release-ceph-luminous conflicts with centos-release-ceph-jewel-1.0-1.el7.centos.noarch"<br>
><br>
<br>
Oops, that was addressed by an upgrade note for the 3.1 release, as<br>
explained here: <a href="https://www.softwarefactory-project.io/releases/3.1/" rel="noreferrer" target="_blank">https://www.softwarefactory-project.io/releases/3.1/</a><br>
(search for "Upgrade Notes")<br>
<br>
You have to do:<br>
"yum remove -y centos-release-ceph-jewel" before installing the<br>
sf-release-3.2.rpm<br>
<br>
Regards,<br>
-Tristan<br>
<br>
> I see that I can fix it adding --skip-broken but.... I am not pretty sure If this is a good idea...I will wait for your opinión. In my first approach I tried to deploy directly sf 3.2 but with our arch I get some errors and at the end I give up. Anyway I can try again.<br>
><br>
><br>
> By other hand, I remount /srv/host-rootfs to try but, this doesn't works.<br>
><br>
><br>
> I attach our arch.yaml in case it could help<br>
><br>
> ________________________________________<br>
> De: Tristan Cacqueray [<a href="mailto:tdecacqu@redhat.com" target="_blank">tdecacqu@redhat.com</a>]<br>
> Enviado: viernes, 12 de abril de 2019 2:55<br>
> Para: Javier Pena; RUIZ LOPEZ Noel; <a href="mailto:softwarefactory-dev@redhat.com" target="_blank">softwarefactory-dev@redhat.com</a><br>
> Asunto: Re: [Softwarefactory-dev] Zuul NODE FAILURE<br>
><br>
> On Thu, Apr 11, 2019 at 12:08 Javier Pena wrote:<br>
>> ----- Original Message -----<br>
>><br>
>>> Hello,<br>
>><br>
>>> I have just deployed software factory and when I tried to test how zuul<br>
>>> works, I got the following zuul error:<br>
>><br>
>>> NODE_FAILURE<br>
>><br>
>>> Now, I can see that nodes always keep "building" state.<br>
>><br>
>>> nodepool log :<br>
>><br>
>>> 2019-04-11 17:55:18,203 ERROR nodepool.NodeLauncher-0000000046: Launch<br>
>>> attempt 9/9 failed for node 0000000046:<br>
>>> Traceback (most recent call last):<br>
>>> File<br>
>>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/oci/handler.py",<br>
>>> line 40, in launch<br>
>>> self.handler.pool, hostid, port, self.label)<br>
>>> File<br>
>>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/oci/provider.py",<br>
>>> line 149, in createContainer<br>
>>> "Manager %s failed to initialized" % <a href="http://self.provider.name" rel="noreferrer" target="_blank">self.provider.name</a>)<br>
>>> RuntimeError: Manager oci-provider-hypervisor-oci failed to initialized<br>
>>> 2019-04-11 17:55:19,208 ERROR nodepool.NodeLauncher-0000000046: Launch failed<br>
>>> for node 0000000046:<br>
>>> Traceback (most recent call last):<br>
>>> File<br>
>>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/__init__.py",<br>
>>> line 659, in run<br>
>>> self.launch()<br>
>>> File<br>
>>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/oci/handler.py",<br>
>>> line 57, in launch<br>
>>> self.node.host_keys = key<br>
>>> UnboundLocalError: local variable 'key' referenced before assignment<br>
>>> 2019-04-11 17:55:19,208 ERROR nodepool.NodeLauncher-0000000045: Launch failed<br>
>>> for node 0000000045:<br>
>>> Traceback (most recent call last):<br>
>>> File<br>
>>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/__init__.py",<br>
>>> line 659, in run<br>
>>> self.launch()<br>
>>> File<br>
>>> "/opt/rh/rh-python35/root/usr/lib/python3.5/site-packages/nodepool/driver/oci/handler.py",<br>
>>> line 57, in launch<br>
>>> self.node.host_keys = key<br>
>>> UnboundLocalError: local variable 'key' referenced before assignment<br>
>>> 2019-04-11 17:55:22,918 INFO nodepool.DeletedNodeWorker: Deleting failed<br>
>>> instance 0000000045-centos-oci-100-0000000045 from<br>
>>> oci-provider-hypervisor-oci<br>
>>> 2019-04-11 17:55:22,926 INFO nodepool.NodeDeleter: Deleting ZK node<br>
>>> id=0000000045, state=deleting,<br>
>>> external_id=0000000045-centos-oci-100-0000000045<br>
>>> 2019-04-11 17:55:22,934 INFO nodepool.DeletedNodeWorker: Deleting failed<br>
>>> instance 0000000046-centos-oci-100-0000000046 from<br>
>>> oci-provider-hypervisor-oci<br>
>>> 2019-04-11 17:55:22,940 INFO nodepool.NodeDeleter: Deleting ZK node<br>
>>> id=0000000046, state=deleting,<br>
>>> external_id=0000000046-centos-oci-100-0000000046<br>
>>> 2019-04-11 17:55:26,276 INFO nodepool.NodePool: Creating requests for 2<br>
>>> centos-oci nodes<br>
>>> 2019-04-11 17:55:29,822 INFO<br>
>>> nodepool.PoolWorker.oci-provider-hypervisor-oci-main: Assigning node request<br>
>>> <NodeRequest {'id': '100-0000000047', 'node_types': ['centos-oci'], 'state':<br>
>>> 'requested', 'state_time': 1554998126.2781763, 'stat':<br>
>>> ZnodeStat(czxid=11466, mzxid=11466, ctime=1554998126279,<br>
>>> mtime=1554998126279, version=0, cversion=0, aversion=0, ephemeralOwner=0,<br>
>>> dataLength=217, numChildren=0, pzxid=11466), 'nodes': [], 'reuse': False,<br>
>>> 'declined_by': [], 'requestor': 'NodePool:min-ready'}><br>
>>> 2019-04-11 17:55:29,845 WARNING nodepool.driver.oci.OpenContainerProvider:<br>
>>> Creating container when provider isn't ready<br>
>><br>
>>> Any idea?<br>
><br>
> Hello Noel,<br>
><br>
> NODE_ERROR indicates a failure to start the nodes, and the exception you<br>
> found in the logs is an issue that has been fixed in newer version.<br>
> It seems like you deployed Software Factory version 3.0, since 3.1 the<br>
> drivers has been renamed runC and greatly improved.<br>
> Can you try to upgrade to version 3.2:<br>
><br>
> <a href="https://www.softwarefactory-project.io/docs/3.2/operator/upgrade.html" rel="noreferrer" target="_blank">https://www.softwarefactory-project.io/docs/3.2/operator/upgrade.html</a><br>
><br>
> After the upgrade process, please restart the instance (that's because<br>
> we don't support upgrade from 3.0, and restart is needed to refresh the<br>
> services).<br>
><br>
><br>
>><br>
>> Hi Noel,<br>
>><br>
>> I'm not sure if it's the same situation, but last time I tried to use oci containers I had to remount /srv/host-rootfs as read-write before it would work the first time (it is mounted as read-only by default). After this and a reboot, it worked fine as ro.<br>
>><br>
>> So can you try a quick "mount -o remount,rw /srv/host-rootfs" and see if it fixes it?<br>
>><br>
><br>
> Since version 3.1 (and the rename to runC), we fixed a critical issue<br>
> with bubblewrap and this remount shouldn't be needed anymore.<br>
><br>
> Regards,<br>
> -Tristan<br>
> description: Minimal Software Factory deployment<br>
> inventory:<br>
> - hostname: <a href="http://managesf.sftests.com" rel="noreferrer" target="_blank">managesf.sftests.com</a><br>
> ip: 10.6.71.81<br>
> name: managesf<br>
> public_url: <a href="https://sftests.com" rel="noreferrer" target="_blank">https://sftests.com</a><br>
> roles:<br>
> - install-server<br>
> - mysql<br>
> - zookeeper<br>
> - gateway<br>
> - cauth<br>
> - managesf<br>
> - etherpad<br>
> - lodgeit<br>
> - gitweb<br>
> - gerrit<br>
> - gerritbot<br>
> - logserver<br>
> - zuul-scheduler<br>
> - zuul-executor<br>
> - zuul-web<br>
> - nodepool-launcher<br>
> - murmur<br>
> - mirror<br>
> - kibana<br>
> - repoxplorer<br>
> - hydrant<br>
> - firehose<br>
> - grafana<br>
> - rabbitmq<br>
> - storyboard<br>
> - storyboard-webclient<br>
> - hostname: <a href="http://elk.sftests.com" rel="noreferrer" target="_blank">elk.sftests.com</a><br>
> ip: 192.168.71.82<br>
> name: elk<br>
> public_url: <a href="http://elk.sftests.com" rel="noreferrer" target="_blank">http://elk.sftests.com</a><br>
> roles:<br>
> - elasticsearch<br>
> - logstash<br>
> - influxdb<br>
> - hostname: <a href="http://nodepool-builder.sftests.com" rel="noreferrer" target="_blank">nodepool-builder.sftests.com</a><br>
> ip: 192.168.71.83<br>
> name: nodepool-builder<br>
> public_url: <a href="http://nodepool-builder.sftests.com" rel="noreferrer" target="_blank">http://nodepool-builder.sftests.com</a><br>
> roles:<br>
> - nodepool-builder<br>
> - hostname: <a href="http://zuul-merger.sftests.com" rel="noreferrer" target="_blank">zuul-merger.sftests.com</a><br>
> ip: 192.168.71.84<br>
> name: zuul-merger<br>
> public_url: <a href="http://zuul-merger.sftests.com" rel="noreferrer" target="_blank">http://zuul-merger.sftests.com</a><br>
> roles:<br>
> - zuul-merger<br>
> - hostname: <a href="http://hypervisor-oci.sftests.com" rel="noreferrer" target="_blank">hypervisor-oci.sftests.com</a><br>
> ip: 192.168.71.86<br>
> max-servers: 10<br>
> name: hypervisor-oci<br>
> public_url: <a href="http://hypervisor-oci.sftests.com" rel="noreferrer" target="_blank">http://hypervisor-oci.sftests.com</a><br>
> remote: true<br>
> roles:<br>
> - hypervisor-oci<br>
_______________________________________________<br>
Softwarefactory-dev mailing list<br>
<a href="mailto:Softwarefactory-dev@redhat.com" target="_blank">Softwarefactory-dev@redhat.com</a><br>
<a href="https://www.redhat.com/mailman/listinfo/softwarefactory-dev" rel="noreferrer" target="_blank">https://www.redhat.com/mailman/listinfo/softwarefactory-dev</a><br>
</blockquote></div>